diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,3823 @@ +{ + "best_metric": 76.93520799124248, + "best_model_checkpoint": "/mounts/work/faeze/attempt/new_setting_outputs/source_adapter/crisis_13/0/checkpoint-1152", + "epoch": 200.0, + "global_step": 3200, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 1.0, + "learning_rate": 1.4999999999999999e-05, + "loss": 9.5194, + "step": 16 + }, + { + "epoch": 1.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Shi\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 9.631518363952637, + "eval_runtime": 2.1772, + "eval_samples_per_second": 53.278, + "step": 16 + }, + { + "epoch": 2.0, + "learning_rate": 2.9999999999999997e-05, + "loss": 8.9553, + "step": 32 + }, + { + "epoch": 2.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 9.292896270751953, + "eval_runtime": 2.1118, + "eval_samples_per_second": 54.93, + "step": 32 + }, + { + "epoch": 3.0, + "learning_rate": 4.4999999999999996e-05, + "loss": 8.2015, + "step": 48 + }, + { + "epoch": 3.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a boat\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 8.517956733703613, + "eval_runtime": 2.0756, + "eval_samples_per_second": 55.889, + "step": 48 + }, + { + "epoch": 4.0, + "learning_rate": 5.9999999999999995e-05, + "loss": 7.0499, + "step": 64 + }, + { + "epoch": 4.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"116.\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@hara go\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Please show us\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Sarah\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification: D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"i will\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 6.531414031982422, + "eval_runtime": 2.4081, + "eval_samples_per_second": 48.171, + "step": 64 + }, + { + "epoch": 5.0, + "learning_rate": 7.5e-05, + "loss": 5.3806, + "step": 80 + }, + { + "epoch": 5.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"1\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"33000f\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Si\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@hara go\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australia\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"CNN\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"COVID\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"FEAR\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"FGS\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Florida\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"France\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"India\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"KGB\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"O U C\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"US\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a grand\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"classification\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"conservative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"dal\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"i\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"jai hin\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"microwave\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_d\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"to irri\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"trud\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 116.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 3.8640472888946533, + "eval_runtime": 2.1363, + "eval_samples_per_second": 54.3, + "step": 80 + }, + { + "epoch": 6.0, + "learning_rate": 8.999999999999999e-05, + "loss": 3.0014, + "step": 96 + }, + { + "epoch": 6.0, + "eval_accuracy": 12.068965517241379, + "eval_average_metrics": 10.466996817571529, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.30434782608695654, \"recall\": 0.45161290322580644, \"f1-score\": 0.3636363636363636, \"support\": 31.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 37.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.1206896551724138, \"macro avg\": {\"precision\": 0.023411371237458196, \"recall\": 0.034739454094292806, \"f1-score\": 0.02797202797202797, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.08133433283358321, \"recall\": 0.1206896551724138, \"f1-score\": 0.09717868338557993, \"support\": 116.0}}", + "eval_f1_macro": 2.7972027972027966, + "eval_f1_micro": 17.28395061728395, + "eval_f1_weighted": 9.717868338557993, + "eval_loss": 1.5544441938400269, + "eval_runtime": 2.0085, + "eval_samples_per_second": 57.753, + "step": 96 + }, + { + "epoch": 7.0, + "learning_rate": 0.00010499999999999999, + "loss": 1.4145, + "step": 112 + }, + { + "epoch": 7.0, + "eval_accuracy": 31.896551724137932, + "eval_average_metrics": 20.735164092162066, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.31896551724137934, \"recall\": 1.0, \"f1-score\": 0.48366013071895425, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.31896551724137934, \"macro avg\": {\"precision\": 0.03987068965517242, \"recall\": 0.125, \"f1-score\": 0.06045751633986928, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.10173900118906065, \"recall\": 0.31896551724137934, \"f1-score\": 0.15427090376380437, \"support\": 116.0}}", + "eval_f1_macro": 3.720462543991956, + "eval_f1_micro": 31.896551724137932, + "eval_f1_weighted": 15.427090376380438, + "eval_loss": 0.8676283359527588, + "eval_runtime": 2.1145, + "eval_samples_per_second": 54.861, + "step": 112 + }, + { + "epoch": 8.0, + "learning_rate": 0.00011999999999999999, + "loss": 0.938, + "step": 128 + }, + { + "epoch": 8.0, + "eval_accuracy": 27.586206896551722, + "eval_average_metrics": 18.972405875133617, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.27184466019417475, \"recall\": 0.9032258064516129, \"f1-score\": 0.417910447761194, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.3076923076923077, \"recall\": 0.10810810810810811, \"f1-score\": 0.16, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.27586206896551724, \"macro avg\": {\"precision\": 0.07244212098581031, \"recall\": 0.1264167393199651, \"f1-score\": 0.07223880597014926, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.17079137802271382, \"recall\": 0.27586206896551724, \"f1-score\": 0.16271744724652598, \"support\": 116.0}}", + "eval_f1_macro": 4.445464982778416, + "eval_f1_micro": 27.586206896551722, + "eval_f1_weighted": 16.2717447246526, + "eval_loss": 0.7047381401062012, + "eval_runtime": 2.1706, + "eval_samples_per_second": 53.441, + "step": 128 + }, + { + "epoch": 9.0, + "learning_rate": 0.000135, + "loss": 0.8323, + "step": 144 + }, + { + "epoch": 9.0, + "eval_accuracy": 37.93103448275862, + "eval_average_metrics": 27.735665281304225, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.36764705882352944, \"recall\": 0.8064516129032258, \"f1-score\": 0.5050505050505051, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.3958333333333333, \"recall\": 0.5135135135135135, \"f1-score\": 0.44705882352941173, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3793103448275862, \"macro avg\": {\"precision\": 0.09543504901960784, \"recall\": 0.1649956408020924, \"f1-score\": 0.1190136660724896, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.22450769100743745, \"recall\": 0.3793103448275862, \"f1-score\": 0.27756674247546453, \"support\": 116.0}}", + "eval_f1_macro": 7.323917912153206, + "eval_f1_micro": 37.93103448275862, + "eval_f1_weighted": 27.756674247546453, + "eval_loss": 0.6856001615524292, + "eval_runtime": 2.0791, + "eval_samples_per_second": 55.794, + "step": 144 + }, + { + "epoch": 10.0, + "learning_rate": 0.00015, + "loss": 0.736, + "step": 160 + }, + { + "epoch": 10.0, + "eval_accuracy": 36.206896551724135, + "eval_average_metrics": 26.25442086648983, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.345679012345679, \"recall\": 0.9032258064516129, \"f1-score\": 0.5, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.4, \"recall\": 0.3783783783783784, \"f1-score\": 0.3888888888888889, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3620689655172414, \"macro avg\": {\"precision\": 0.09320987654320988, \"recall\": 0.1602005231037489, \"f1-score\": 0.1111111111111111, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.21996594295444868, \"recall\": 0.3620689655172414, \"f1-score\": 0.25766283524904215, \"support\": 116.0}}", + "eval_f1_macro": 6.837606837606837, + "eval_f1_micro": 36.20689655172414, + "eval_f1_weighted": 25.766283524904214, + "eval_loss": 0.6892978549003601, + "eval_runtime": 2.0856, + "eval_samples_per_second": 55.619, + "step": 160 + }, + { + "epoch": 11.0, + "learning_rate": 0.000165, + "loss": 0.7248, + "step": 176 + }, + { + "epoch": 11.0, + "eval_accuracy": 39.6551724137931, + "eval_average_metrics": 29.085576143629396, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.4166666666666667, \"recall\": 0.6451612903225806, \"f1-score\": 0.5063291139240507, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.38235294117647056, \"recall\": 0.7027027027027027, \"f1-score\": 0.4952380952380953, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.39655172413793105, \"macro avg\": {\"precision\": 0.09987745098039216, \"recall\": 0.16848299912816042, \"f1-score\": 0.12519590114526824, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.2333079783637593, \"recall\": 0.39655172413793105, \"f1-score\": 0.2932759659953026, \"support\": 116.0}}", + "eval_f1_macro": 7.704363147401122, + "eval_f1_micro": 39.6551724137931, + "eval_f1_weighted": 29.32759659953026, + "eval_loss": 0.6886057257652283, + "eval_runtime": 2.1055, + "eval_samples_per_second": 55.093, + "step": 176 + }, + { + "epoch": 12.0, + "learning_rate": 0.00017999999999999998, + "loss": 0.6738, + "step": 192 + }, + { + "epoch": 12.0, + "eval_accuracy": 37.93103448275862, + "eval_average_metrics": 27.760615065515246, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.41025641025641024, \"recall\": 0.5161290322580645, \"f1-score\": 0.4571428571428572, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.36363636363636365, \"recall\": 0.7567567567567568, \"f1-score\": 0.4912280701754386, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3793103448275862, \"macro avg\": {\"precision\": 0.09673659673659674, \"recall\": 0.15911072362685266, \"f1-score\": 0.11854636591478698, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.2256249497628808, \"recall\": 0.3793103448275862, \"f1-score\": 0.27885230317172244, \"support\": 116.0}}", + "eval_f1_macro": 7.295160979371507, + "eval_f1_micro": 37.93103448275862, + "eval_f1_weighted": 27.885230317172244, + "eval_loss": 0.6640838980674744, + "eval_runtime": 2.2026, + "eval_samples_per_second": 52.665, + "step": 192 + }, + { + "epoch": 13.0, + "learning_rate": 0.000195, + "loss": 0.7337, + "step": 208 + }, + { + "epoch": 13.0, + "eval_accuracy": 37.06896551724138, + "eval_average_metrics": 27.156620348964545, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.38636363636363635, \"recall\": 0.5483870967741935, \"f1-score\": 0.4533333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.3611111111111111, \"recall\": 0.7027027027027027, \"f1-score\": 0.47706422018348627, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3706896551724138, \"macro avg\": {\"precision\": 0.09343434343434343, \"recall\": 0.15638622493461202, \"f1-score\": 0.11629969418960245, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.21843434343434343, \"recall\": 0.3706896551724138, \"f1-score\": 0.2733164610355373, \"support\": 116.0}}", + "eval_f1_macro": 7.156904257821689, + "eval_f1_micro": 37.06896551724138, + "eval_f1_weighted": 27.33164610355373, + "eval_loss": 0.6635391712188721, + "eval_runtime": 2.234, + "eval_samples_per_second": 51.925, + "step": 208 + }, + { + "epoch": 14.0, + "learning_rate": 0.00020999999999999998, + "loss": 0.7431, + "step": 224 + }, + { + "epoch": 14.0, + "eval_accuracy": 37.93103448275862, + "eval_average_metrics": 27.575873831071966, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.46153846153846156, \"recall\": 0.3870967741935484, \"f1-score\": 0.42105263157894735, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.35555555555555557, \"recall\": 0.8648648648648649, \"f1-score\": 0.5039370078740157, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3793103448275862, \"macro avg\": {\"precision\": 0.10213675213675213, \"recall\": 0.15649520488230168, \"f1-score\": 0.11562370493162039, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.23675213675213677, \"recall\": 0.3793103448275862, \"f1-score\": 0.2732612143990168, \"support\": 116.0}}", + "eval_f1_macro": 7.115304918868946, + "eval_f1_micro": 37.93103448275862, + "eval_f1_weighted": 27.32612143990168, + "eval_loss": 0.6511205434799194, + "eval_runtime": 2.2897, + "eval_samples_per_second": 50.662, + "step": 224 + }, + { + "epoch": 15.0, + "learning_rate": 0.000225, + "loss": 0.7529, + "step": 240 + }, + { + "epoch": 15.0, + "eval_accuracy": 38.793103448275865, + "eval_average_metrics": 28.441580913172167, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.39285714285714285, \"recall\": 0.7096774193548387, \"f1-score\": 0.5057471264367817, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.38333333333333336, \"recall\": 0.6216216216216216, \"f1-score\": 0.4742268041237113, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3879310344827586, \"macro avg\": {\"precision\": 0.09702380952380953, \"recall\": 0.16641238012205756, \"f1-score\": 0.12249674132006162, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.22725779967159276, \"recall\": 0.3879310344827586, \"f1-score\": 0.28641855751825473, \"support\": 116.0}}", + "eval_f1_macro": 7.538261004311485, + "eval_f1_micro": 38.793103448275865, + "eval_f1_weighted": 28.641855751825474, + "eval_loss": 0.6641126871109009, + "eval_runtime": 2.221, + "eval_samples_per_second": 52.229, + "step": 240 + }, + { + "epoch": 16.0, + "learning_rate": 0.00023999999999999998, + "loss": 0.7239, + "step": 256 + }, + { + "epoch": 16.0, + "eval_accuracy": 36.206896551724135, + "eval_average_metrics": 26.256044568490083, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.4230769230769231, \"recall\": 0.3548387096774194, \"f1-score\": 0.3859649122807018, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.34444444444444444, \"recall\": 0.8378378378378378, \"f1-score\": 0.4881889763779528, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3620689655172414, \"macro avg\": {\"precision\": 0.09594017094017093, \"recall\": 0.14908456843940715, \"f1-score\": 0.10926923608233183, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.22292956086059534, \"recall\": 0.3620689655172414, \"f1-score\": 0.25886124488522416, \"support\": 116.0}}", + "eval_f1_macro": 6.724260681989651, + "eval_f1_micro": 36.20689655172414, + "eval_f1_weighted": 25.886124488522416, + "eval_loss": 0.6616472601890564, + "eval_runtime": 2.2101, + "eval_samples_per_second": 52.487, + "step": 256 + }, + { + "epoch": 17.0, + "learning_rate": 0.00025499999999999996, + "loss": 0.7121, + "step": 272 + }, + { + "epoch": 17.0, + "eval_accuracy": 38.793103448275865, + "eval_average_metrics": 28.220869647593787, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.4057971014492754, \"recall\": 0.9032258064516129, \"f1-score\": 0.5599999999999999, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.3617021276595745, \"recall\": 0.4594594594594595, \"f1-score\": 0.40476190476190477, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.3879310344827586, \"macro avg\": {\"precision\": 0.09593740363860623, \"recall\": 0.17033565823888405, \"f1-score\": 0.12059523809523809, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.22381628334768788, \"recall\": 0.3879310344827586, \"f1-score\": 0.27876026272578, \"support\": 116.0}}", + "eval_f1_macro": 7.421245421245421, + "eval_f1_micro": 38.793103448275865, + "eval_f1_weighted": 27.876026272578, + "eval_loss": 0.6310274600982666, + "eval_runtime": 2.2051, + "eval_samples_per_second": 52.605, + "step": 272 + }, + { + "epoch": 18.0, + "learning_rate": 0.00027, + "loss": 0.6385, + "step": 288 + }, + { + "epoch": 18.0, + "eval_accuracy": 42.241379310344826, + "eval_average_metrics": 31.05062473823817, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 8.0}, \"challenge\": {\"precision\": 0.4888888888888889, \"recall\": 0.7096774193548387, \"f1-score\": 0.5789473684210527, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.38028169014084506, \"recall\": 0.7297297297297297, \"f1-score\": 0.5, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.4224137931034483, \"macro avg\": {\"precision\": 0.10864632237871674, \"recall\": 0.17992589363557104, \"f1-score\": 0.13486842105263158, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.25194808698936916, \"recall\": 0.4224137931034483, \"f1-score\": 0.3142014519056261, \"support\": 116.0}}", + "eval_f1_macro": 8.299595141700406, + "eval_f1_micro": 42.24137931034483, + "eval_f1_weighted": 31.42014519056261, + "eval_loss": 0.6071025133132935, + "eval_runtime": 2.27, + "eval_samples_per_second": 51.101, + "step": 288 + }, + { + "epoch": 19.0, + "learning_rate": 0.000285, + "loss": 0.5922, + "step": 304 + }, + { + "epoch": 19.0, + "eval_accuracy": 45.689655172413794, + "eval_average_metrics": 36.16077557122249, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.25, \"recall\": 0.125, \"f1-score\": 0.16666666666666666, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5510204081632653, \"recall\": 0.8709677419354839, \"f1-score\": 0.6749999999999999, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.1, \"recall\": 0.1111111111111111, \"f1-score\": 0.10526315789473685, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.44680851063829785, \"recall\": 0.5675675675675675, \"f1-score\": 0.5, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.3333333333333333, \"f1-score\": 0.4, \"support\": 9.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"accuracy\": 0.45689655172413796, \"macro avg\": {\"precision\": 0.23097861485019539, \"recall\": 0.25099746924343697, \"f1-score\": 0.23086622807017543, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.353565065057571, \"recall\": 0.45689655172413796, \"f1-score\": 0.39056639443436175, \"support\": 116.0}}", + "eval_f1_macro": 14.207152496626179, + "eval_f1_micro": 45.689655172413794, + "eval_f1_weighted": 39.056639443436175, + "eval_loss": 0.5678648352622986, + "eval_runtime": 2.2264, + "eval_samples_per_second": 52.103, + "step": 304 + }, + { + "epoch": 20.0, + "learning_rate": 0.0003, + "loss": 0.6507, + "step": 320 + }, + { + "epoch": 20.0, + "eval_accuracy": 50.0, + "eval_average_metrics": 40.86473713418625, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.375, \"recall\": 0.25, \"f1-score\": 0.3, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.5, \"recall\": 0.375, \"f1-score\": 0.42857142857142855, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5094339622641509, \"recall\": 0.8709677419354839, \"f1-score\": 0.6428571428571428, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.5, \"recall\": 0.1111111111111111, \"f1-score\": 0.1818181818181818, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.4888888888888889, \"recall\": 0.5945945945945946, \"f1-score\": 0.5365853658536586, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.2222222222222222, \"f1-score\": 0.3636363636363636, \"support\": 9.0}, \"accuracy\": 0.5, \"macro avg\": {\"precision\": 0.42166535639413, \"recall\": 0.30298695873292647, \"f1-score\": 0.3066835603420969, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.4817357044748066, \"recall\": 0.5, \"f1-score\": 0.44586114054154424, \"support\": 116.0}}", + "eval_f1_macro": 18.87283448259058, + "eval_f1_micro": 50.0, + "eval_f1_weighted": 44.58611405415442, + "eval_loss": 0.5527870655059814, + "eval_runtime": 2.2006, + "eval_samples_per_second": 52.713, + "step": 320 + }, + { + "epoch": 21.0, + "learning_rate": 0.00029833333333333334, + "loss": 0.6159, + "step": 336 + }, + { + "epoch": 21.0, + "eval_accuracy": 50.0, + "eval_average_metrics": 39.89382649262413, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.375, \"f1-score\": 0.5, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5306122448979592, \"recall\": 0.8387096774193549, \"f1-score\": 0.65, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.1111111111111111, \"f1-score\": 0.19999999999999998, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.4482758620689655, \"recall\": 0.7027027027027027, \"f1-score\": 0.5473684210526315, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.2222222222222222, \"f1-score\": 0.3333333333333333, \"support\": 9.0}, \"accuracy\": 0.5, \"macro avg\": {\"precision\": 0.42444434670419895, \"recall\": 0.28121821418192383, \"f1-score\": 0.27883771929824563, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.4658205731757626, \"recall\": 0.5, \"f1-score\": 0.4241606170598911, \"support\": 116.0}}", + "eval_f1_macro": 17.15924426450742, + "eval_f1_micro": 50.0, + "eval_f1_weighted": 42.41606170598911, + "eval_loss": 0.49902039766311646, + "eval_runtime": 2.1471, + "eval_samples_per_second": 54.026, + "step": 336 + }, + { + "epoch": 22.0, + "learning_rate": 0.00029666666666666665, + "loss": 0.4905, + "step": 352 + }, + { + "epoch": 22.0, + "eval_accuracy": 55.172413793103445, + "eval_average_metrics": 47.567859823700545, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.3, \"recall\": 0.5, \"f1-score\": 0.37499999999999994, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.5, \"f1-score\": 0.5714285714285715, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.4444444444444444, \"f1-score\": 0.6153846153846153, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.49019607843137253, \"recall\": 0.6756756756756757, \"f1-score\": 0.5681818181818181, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"request\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"accuracy\": 0.5517241379310345, \"macro avg\": {\"precision\": 0.4778254357298475, \"recall\": 0.41510462074978205, \"f1-score\": 0.4259856839408056, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.5574013034332508, \"recall\": 0.5517241379310345, \"f1-score\": 0.5371210808146878, \"support\": 116.0}}", + "eval_f1_macro": 26.2145036271265, + "eval_f1_micro": 55.172413793103445, + "eval_f1_weighted": 53.71210808146878, + "eval_loss": 0.4539774954319, + "eval_runtime": 2.1666, + "eval_samples_per_second": 53.539, + "step": 352 + }, + { + "epoch": 23.0, + "learning_rate": 0.00029499999999999996, + "loss": 0.4571, + "step": 368 + }, + { + "epoch": 23.0, + "eval_accuracy": 57.758620689655174, + "eval_average_metrics": 51.260575594623866, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.4166666666666667, \"recall\": 0.8333333333333334, \"f1-score\": 0.5555555555555556, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8333333333333334, \"recall\": 0.3225806451612903, \"f1-score\": 0.4651162790697674, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8571428571428571, \"recall\": 0.6666666666666666, \"f1-score\": 0.75, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6, \"recall\": 0.6486486486486487, \"f1-score\": 0.6233766233766234, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.35, \"recall\": 0.7777777777777778, \"f1-score\": 0.48275862068965514, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.4444444444444444, \"f1-score\": 0.5714285714285714, \"support\": 9.0}, \"accuracy\": 0.5775862068965517, \"macro avg\": {\"precision\": 0.5758928571428572, \"recall\": 0.5554314395040202, \"f1-score\": 0.5247794562650215, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6646346469622332, \"recall\": 0.5775862068965517, \"f1-score\": 0.572309406136453, \"support\": 116.0}}", + "eval_f1_macro": 32.29412038553979, + "eval_f1_micro": 57.758620689655174, + "eval_f1_weighted": 57.23094061364532, + "eval_loss": 0.4593088924884796, + "eval_runtime": 2.1213, + "eval_samples_per_second": 54.683, + "step": 368 + }, + { + "epoch": 24.0, + "learning_rate": 0.00029333333333333327, + "loss": 0.3941, + "step": 384 + }, + { + "epoch": 24.0, + "eval_accuracy": 66.37931034482759, + "eval_average_metrics": 58.64231561905229, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.45454545454545453, \"recall\": 0.8333333333333334, \"f1-score\": 0.5882352941176471, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.7419354838709677, \"f1-score\": 0.7419354838709677, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.6666666666666666, \"f1-score\": 0.8, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.6486486486486487, \"f1-score\": 0.6575342465753425, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.2857142857142857, \"recall\": 0.2222222222222222, \"f1-score\": 0.25, \"support\": 9.0}, \"request\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"accuracy\": 0.6637931034482759, \"macro avg\": {\"precision\": 0.6123577363496717, \"recall\": 0.5661841276760632, \"f1-score\": 0.5750912793309989, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6848503507986266, \"recall\": 0.6637931034482759, \"f1-score\": 0.6642040921233869, \"support\": 116.0}}", + "eval_f1_macro": 35.39023257421532, + "eval_f1_micro": 66.37931034482759, + "eval_f1_weighted": 66.4204092123387, + "eval_loss": 0.388798326253891, + "eval_runtime": 2.0824, + "eval_samples_per_second": 55.705, + "step": 384 + }, + { + "epoch": 25.0, + "learning_rate": 0.00029166666666666664, + "loss": 0.3707, + "step": 400 + }, + { + "epoch": 25.0, + "eval_accuracy": 60.3448275862069, + "eval_average_metrics": 53.652705233850234, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.3333333333333333, \"recall\": 0.875, \"f1-score\": 0.48275862068965514, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5806451612903226, \"recall\": 0.5806451612903226, \"f1-score\": 0.5806451612903226, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.782608695652174, \"recall\": 0.4864864864864865, \"f1-score\": 0.6000000000000001, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.42857142857142855, \"recall\": 0.3333333333333333, \"f1-score\": 0.375, \"support\": 9.0}, \"request\": {\"precision\": 0.8571428571428571, \"recall\": 0.6666666666666666, \"f1-score\": 0.75, \"support\": 9.0}, \"accuracy\": 0.603448275862069, \"macro avg\": {\"precision\": 0.5533432400543199, \"recall\": 0.5830442337498789, \"f1-score\": 0.5429949171919417, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6539765831370029, \"recall\": 0.603448275862069, \"f1-score\": 0.6050609393579073, \"support\": 116.0}}", + "eval_f1_macro": 33.415071827196414, + "eval_f1_micro": 60.3448275862069, + "eval_f1_weighted": 60.50609393579073, + "eval_loss": 0.4242806136608124, + "eval_runtime": 2.2084, + "eval_samples_per_second": 52.527, + "step": 400 + }, + { + "epoch": 26.0, + "learning_rate": 0.00029, + "loss": 0.339, + "step": 416 + }, + { + "epoch": 26.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 61.77302749117365, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.6666666666666666, \"f1-score\": 0.8, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.7297297297297297, \"f1-score\": 0.7105263157894737, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.4444444444444444, \"f1-score\": 0.47058823529411764, \"support\": 9.0}, \"request\": {\"precision\": 0.8333333333333334, \"recall\": 0.5555555555555556, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.6634729853479853, \"recall\": 0.5902081517000872, \"f1-score\": 0.6145224357685658, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7110932171276998, \"recall\": 0.6982758620689655, \"f1-score\": 0.6962017227283588, \"support\": 116.0}}", + "eval_f1_macro": 37.81676527806559, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.62017227283587, + "eval_loss": 0.3623848557472229, + "eval_runtime": 2.1161, + "eval_samples_per_second": 54.817, + "step": 416 + }, + { + "epoch": 27.0, + "learning_rate": 0.0002883333333333333, + "loss": 0.3184, + "step": 432 + }, + { + "epoch": 27.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 62.719846550445325, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.875, \"f1-score\": 0.9333333333333333, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.4444444444444444, \"recall\": 0.4444444444444444, \"f1-score\": 0.4444444444444444, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.6264456711325782, \"recall\": 0.657731824566502, \"f1-score\": 0.6369607776174497, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7091025084318285, \"recall\": 0.7068965517241379, \"f1-score\": 0.7030248954203372, \"support\": 116.0}}", + "eval_f1_macro": 39.19758631491999, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.30248954203373, + "eval_loss": 0.3541596531867981, + "eval_runtime": 2.1012, + "eval_samples_per_second": 55.205, + "step": 432 + }, + { + "epoch": 28.0, + "learning_rate": 0.0002866666666666667, + "loss": 0.2422, + "step": 448 + }, + { + "epoch": 28.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 63.936806022349785, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6857142857142857, \"recall\": 0.7741935483870968, \"f1-score\": 0.7272727272727272, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.6666666666666666, \"f1-score\": 0.8, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6590909090909091, \"recall\": 0.7837837837837838, \"f1-score\": 0.7160493827160493, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7028228715728716, \"recall\": 0.6044693887435824, \"f1-score\": 0.6362226508059842, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7361509180474698, \"recall\": 0.7241379310344828, \"f1-score\": 0.717674747559805, \"support\": 116.0}}", + "eval_f1_macro": 39.152163126522105, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 71.7674747559805, + "eval_loss": 0.3609805405139923, + "eval_runtime": 2.2843, + "eval_samples_per_second": 50.782, + "step": 448 + }, + { + "epoch": 29.0, + "learning_rate": 0.000285, + "loss": 0.2425, + "step": 464 + }, + { + "epoch": 29.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 62.08778783554499, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.6159519952866727, \"recall\": 0.6775389300590913, \"f1-score\": 0.6382293836722224, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7044631963430629, \"recall\": 0.6982758620689655, \"f1-score\": 0.6942032454855775, \"support\": 116.0}}", + "eval_f1_macro": 39.275654379829064, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.42032454855776, + "eval_loss": 0.36870959401130676, + "eval_runtime": 2.0508, + "eval_samples_per_second": 56.564, + "step": 464 + }, + { + "epoch": 30.0, + "learning_rate": 0.0002833333333333333, + "loss": 0.1556, + "step": 480 + }, + { + "epoch": 30.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.72567209776729, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7692307692307693, \"recall\": 0.5405405405405406, \"f1-score\": 0.6349206349206349, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.742926726051726, \"recall\": 0.8264721374600407, \"f1-score\": 0.7692301508877595, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7302372340303375, \"recall\": 0.7068965517241379, \"f1-score\": 0.7018613799161024, \"support\": 116.0}}", + "eval_f1_macro": 47.33724005463135, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.18613799161024, + "eval_loss": 0.409758597612381, + "eval_runtime": 2.1135, + "eval_samples_per_second": 54.884, + "step": 480 + }, + { + "epoch": 31.0, + "learning_rate": 0.00028166666666666666, + "loss": 0.1426, + "step": 496 + }, + { + "epoch": 31.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.17670734030372, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7652417027417027, \"recall\": 0.7837202121476314, \"f1-score\": 0.7695496876911684, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7347100935463005, \"recall\": 0.7241379310344828, \"f1-score\": 0.7252233929640028, \"support\": 116.0}}", + "eval_f1_macro": 47.356903857918056, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.52233929640029, + "eval_loss": 0.38606610894203186, + "eval_runtime": 2.0992, + "eval_samples_per_second": 55.258, + "step": 496 + }, + { + "epoch": 32.0, + "learning_rate": 0.00028, + "loss": 0.1403, + "step": 512 + }, + { + "epoch": 32.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.83586644295046, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5, \"recall\": 0.8333333333333334, \"f1-score\": 0.625, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.6129032258064516, \"f1-score\": 0.6551724137931035, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6486486486486487, \"f1-score\": 0.7058823529411764, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7552232676224612, \"recall\": 0.7879023176402209, \"f1-score\": 0.7647955885318435, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7255170353891154, \"recall\": 0.7068965517241379, \"f1-score\": 0.7089981151732236, \"support\": 116.0}}", + "eval_f1_macro": 47.0643439096519, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.89981151732236, + "eval_loss": 0.4265669584274292, + "eval_runtime": 2.1308, + "eval_samples_per_second": 54.439, + "step": 512 + }, + { + "epoch": 33.0, + "learning_rate": 0.00027833333333333334, + "loss": 0.1089, + "step": 528 + }, + { + "epoch": 33.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.9963380759733, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.875, \"recall\": 0.875, \"f1-score\": 0.875, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8571428571428571, \"recall\": 0.5806451612903226, \"f1-score\": 0.6923076923076923, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7712593510290878, \"recall\": 0.8078940835997288, \"f1-score\": 0.7801354873645527, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7594117543073986, \"recall\": 0.7327586206896551, \"f1-score\": 0.7342529048198969, \"support\": 116.0}}", + "eval_f1_macro": 48.00833768397247, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.4252904819897, + "eval_loss": 0.5401732325553894, + "eval_runtime": 2.2438, + "eval_samples_per_second": 51.697, + "step": 528 + }, + { + "epoch": 34.0, + "learning_rate": 0.00027666666666666665, + "loss": 0.1282, + "step": 544 + }, + { + "epoch": 34.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.73343394035153, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.55, \"recall\": 0.9166666666666666, \"f1-score\": 0.6874999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7600610378183907, \"recall\": 0.7728373534825147, \"f1-score\": 0.7583272391919456, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7259054713035443, \"recall\": 0.7068965517241379, \"f1-score\": 0.7088813377399731, \"support\": 116.0}}", + "eval_f1_macro": 46.666291642581264, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.88813377399731, + "eval_loss": 0.5074975490570068, + "eval_runtime": 2.3982, + "eval_samples_per_second": 48.37, + "step": 544 + }, + { + "epoch": 35.0, + "learning_rate": 0.00027499999999999996, + "loss": 0.113, + "step": 560 + }, + { + "epoch": 35.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0757956863876, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6857142857142857, \"recall\": 0.7741935483870968, \"f1-score\": 0.7272727272727272, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"request\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.8116834366834367, \"recall\": 0.7644731424973361, \"f1-score\": 0.7833908982721912, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7446932377966862, \"recall\": 0.7327586206896551, \"f1-score\": 0.7354278794471526, \"support\": 116.0}}", + "eval_f1_macro": 48.208670662904076, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.54278794471526, + "eval_loss": 0.47861307859420776, + "eval_runtime": 2.0863, + "eval_samples_per_second": 55.601, + "step": 560 + }, + { + "epoch": 36.0, + "learning_rate": 0.00027333333333333333, + "loss": 0.0821, + "step": 576 + }, + { + "epoch": 36.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.40700236389213, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5333333333333333, \"recall\": 0.8888888888888888, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7447649572649573, \"recall\": 0.8007089751041363, \"f1-score\": 0.7626605641383726, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7350243147656942, \"recall\": 0.7155172413793104, \"f1-score\": 0.715916033865759, \"support\": 116.0}}", + "eval_f1_macro": 46.93295779313061, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.59160338657588, + "eval_loss": 0.542853057384491, + "eval_runtime": 2.0928, + "eval_samples_per_second": 55.428, + "step": 576 + }, + { + "epoch": 37.0, + "learning_rate": 0.00027166666666666664, + "loss": 0.0757, + "step": 592 + }, + { + "epoch": 37.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.66949561846265, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7844065656565656, \"recall\": 0.7921601036520391, \"f1-score\": 0.7787362300689507, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368730407523512, \"recall\": 0.7155172413793104, \"f1-score\": 0.7165230465528387, \"support\": 116.0}}", + "eval_f1_macro": 47.922229542704656, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.65230465528387, + "eval_loss": 0.6507958769798279, + "eval_runtime": 1.9704, + "eval_samples_per_second": 58.871, + "step": 592 + }, + { + "epoch": 38.0, + "learning_rate": 0.00027, + "loss": 0.0588, + "step": 608 + }, + { + "epoch": 38.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.73703365588325, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.631578947368421, \"recall\": 0.7741935483870968, \"f1-score\": 0.6956521739130435, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.3333333333333333, \"f1-score\": 0.4, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7638768039049236, \"recall\": 0.7369768962510898, \"f1-score\": 0.7422196796338673, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7099617264797143, \"recall\": 0.6982758620689655, \"f1-score\": 0.696179050015019, \"support\": 116.0}}", + "eval_f1_macro": 45.67505720823799, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.6179050015019, + "eval_loss": 0.6356076002120972, + "eval_runtime": 2.1143, + "eval_samples_per_second": 54.865, + "step": 608 + }, + { + "epoch": 39.0, + "learning_rate": 0.0002683333333333333, + "loss": 0.109, + "step": 624 + }, + { + "epoch": 39.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.851106421197, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5238095238095238, \"recall\": 0.9166666666666666, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6486486486486487, \"f1-score\": 0.7164179104477612, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7929834054834055, \"recall\": 0.8086796473893247, \"f1-score\": 0.7896230372710594, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7702754142409315, \"recall\": 0.7413793103448276, \"f1-score\": 0.7453637670683421, \"support\": 116.0}}", + "eval_f1_macro": 48.592186908988275, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.5363767068342, + "eval_loss": 0.6028116345405579, + "eval_runtime": 2.1145, + "eval_samples_per_second": 54.859, + "step": 624 + }, + { + "epoch": 40.0, + "learning_rate": 0.0002666666666666666, + "loss": 0.0434, + "step": 640 + }, + { + "epoch": 40.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.73082981916082, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7822721606877204, \"recall\": 0.7946030708127482, \"f1-score\": 0.7866332421998508, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7462762389740077, \"recall\": 0.7413793103448276, \"f1-score\": 0.7423925768768693, \"support\": 116.0}}", + "eval_f1_macro": 48.40819951999081, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.23925768768693, + "eval_loss": 0.5737775564193726, + "eval_runtime": 1.9992, + "eval_samples_per_second": 58.024, + "step": 640 + }, + { + "epoch": 41.0, + "learning_rate": 0.000265, + "loss": 0.0482, + "step": 656 + }, + { + "epoch": 41.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.19735127899169, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.7741935483870968, \"f1-score\": 0.7384615384615385, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7794730392156863, \"recall\": 0.7751713406955343, \"f1-score\": 0.7704721495741973, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7403693373901284, \"recall\": 0.7241379310344828, \"f1-score\": 0.7254814816604266, \"support\": 116.0}}", + "eval_f1_macro": 47.413670743027524, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.54814816604267, + "eval_loss": 0.6634430885314941, + "eval_runtime": 2.0377, + "eval_samples_per_second": 56.927, + "step": 656 + }, + { + "epoch": 42.0, + "learning_rate": 0.0002633333333333333, + "loss": 0.0329, + "step": 672 + }, + { + "epoch": 42.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.16273743305499, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.7027027027027027, \"f1-score\": 0.7323943661971832, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7632253911665676, \"recall\": 0.7830663324614937, \"f1-score\": 0.7668365408904801, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7392959343263603, \"recall\": 0.7241379310344828, \"f1-score\": 0.7263342254744771, \"support\": 116.0}}", + "eval_f1_macro": 47.1899409778757, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.63342254744771, + "eval_loss": 0.6547279357910156, + "eval_runtime": 2.1903, + "eval_samples_per_second": 52.96, + "step": 672 + }, + { + "epoch": 43.0, + "learning_rate": 0.00026166666666666667, + "loss": 0.0316, + "step": 688 + }, + { + "epoch": 43.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 64.21844203108832, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.4583333333333333, \"recall\": 0.9166666666666666, \"f1-score\": 0.611111111111111, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8571428571428571, \"recall\": 0.5806451612903226, \"f1-score\": 0.6923076923076923, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7718700570263071, \"recall\": 0.7891722367528818, \"f1-score\": 0.7608901719195837, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7494119565779048, \"recall\": 0.6982758620689655, \"f1-score\": 0.7039458513089345, \"support\": 116.0}}", + "eval_f1_macro": 46.824010579666684, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 70.39458513089345, + "eval_loss": 0.7738772034645081, + "eval_runtime": 2.0549, + "eval_samples_per_second": 56.451, + "step": 688 + }, + { + "epoch": 44.0, + "learning_rate": 0.00026, + "loss": 0.0505, + "step": 704 + }, + { + "epoch": 44.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.97464242891904, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5333333333333333, \"recall\": 0.6666666666666666, \"f1-score\": 0.5925925925925926, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.739340990550668, \"recall\": 0.7805295214569408, \"f1-score\": 0.7567426750203485, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7020418309906631, \"recall\": 0.6982758620689655, \"f1-score\": 0.6967461730063084, \"support\": 116.0}}", + "eval_f1_macro": 46.56878000125221, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.67461730063084, + "eval_loss": 0.6857355833053589, + "eval_runtime": 2.1516, + "eval_samples_per_second": 53.912, + "step": 704 + }, + { + "epoch": 45.0, + "learning_rate": 0.00025833333333333334, + "loss": 0.0194, + "step": 720 + }, + { + "epoch": 45.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.93408505993239, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.675, \"recall\": 0.7297297297297297, \"f1-score\": 0.7012987012987014, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.8203598484848486, \"recall\": 0.7875647825244599, \"f1-score\": 0.7984059141284494, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7524229362591431, \"recall\": 0.7413793103448276, \"f1-score\": 0.7432780653209022, \"support\": 116.0}}", + "eval_f1_macro": 49.132671638673806, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.32780653209022, + "eval_loss": 0.6950328946113586, + "eval_runtime": 2.1933, + "eval_samples_per_second": 52.888, + "step": 720 + }, + { + "epoch": 46.0, + "learning_rate": 0.00025666666666666665, + "loss": 0.0199, + "step": 736 + }, + { + "epoch": 46.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.64614471265455, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7441835336285052, \"recall\": 0.788127845587523, \"f1-score\": 0.7597323795105688, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7173174442190671, \"recall\": 0.7068965517241379, \"f1-score\": 0.7045250668975559, \"support\": 116.0}}", + "eval_f1_macro": 46.752761816035, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.4525066897556, + "eval_loss": 0.8003849983215332, + "eval_runtime": 2.0532, + "eval_samples_per_second": 56.498, + "step": 736 + }, + { + "epoch": 47.0, + "learning_rate": 0.00025499999999999996, + "loss": 0.0231, + "step": 752 + }, + { + "epoch": 47.0, + "eval_accuracy": 65.51724137931035, + "eval_average_metrics": 60.04096842765952, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.5641025641025641, \"recall\": 0.7096774193548387, \"f1-score\": 0.6285714285714286, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.72, \"recall\": 0.4864864864864865, \"f1-score\": 0.5806451612903226, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5714285714285714, \"recall\": 0.4444444444444444, \"f1-score\": 0.5, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6551724137931034, \"macro avg\": {\"precision\": 0.7158699633699634, \"recall\": 0.7397982660079434, \"f1-score\": 0.7183717533665861, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6690766704559807, \"recall\": 0.6551724137931034, \"f1-score\": 0.6492189843715057, \"support\": 116.0}}", + "eval_f1_macro": 44.207492514866836, + "eval_f1_micro": 65.51724137931035, + "eval_f1_weighted": 64.92189843715057, + "eval_loss": 0.827684760093689, + "eval_runtime": 2.1386, + "eval_samples_per_second": 54.241, + "step": 752 + }, + { + "epoch": 48.0, + "learning_rate": 0.00025333333333333333, + "loss": 0.0259, + "step": 768 + }, + { + "epoch": 48.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.95719742042525, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.811489898989899, \"recall\": 0.8009874794148988, \"f1-score\": 0.7971726856406487, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7623040752351098, \"recall\": 0.7413793103448276, \"f1-score\": 0.744961469579263, \"support\": 116.0}}", + "eval_f1_macro": 49.05678065480915, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.4961469579263, + "eval_loss": 0.7849622964859009, + "eval_runtime": 2.115, + "eval_samples_per_second": 54.847, + "step": 768 + }, + { + "epoch": 49.0, + "learning_rate": 0.00025166666666666664, + "loss": 0.0132, + "step": 784 + }, + { + "epoch": 49.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.94372313914248, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.7741935483870968, \"f1-score\": 0.7500000000000001, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6896551724137931, \"recall\": 0.5405405405405406, \"f1-score\": 0.6060606060606061, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7684892523538168, \"recall\": 0.7650362055603992, \"f1-score\": 0.7542708615704347, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7174182617713971, \"recall\": 0.6982758620689655, \"f1-score\": 0.6970305173844238, \"support\": 116.0}}", + "eval_f1_macro": 46.41666840433444, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.70305173844238, + "eval_loss": 0.8554142117500305, + "eval_runtime": 2.2913, + "eval_samples_per_second": 50.627, + "step": 784 + }, + { + "epoch": 50.0, + "learning_rate": 0.00025, + "loss": 0.0138, + "step": 800 + }, + { + "epoch": 50.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.15480831429107, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.7419354838709677, \"f1-score\": 0.7301587301587302, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7575394403250006, \"recall\": 0.7990107042526398, \"f1-score\": 0.7728733766233766, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7321003628505858, \"recall\": 0.7241379310344828, \"f1-score\": 0.7223020848882918, \"support\": 116.0}}", + "eval_f1_macro": 47.561438561438564, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.23020848882918, + "eval_loss": 0.7713599801063538, + "eval_runtime": 2.2362, + "eval_samples_per_second": 51.874, + "step": 800 + }, + { + "epoch": 51.0, + "learning_rate": 0.0002483333333333333, + "loss": 0.0306, + "step": 816 + }, + { + "epoch": 51.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.99442064338926, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.84, \"recall\": 0.6774193548387096, \"f1-score\": 0.75, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7696143247904708, \"recall\": 0.815157899835319, \"f1-score\": 0.7817010406313063, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.754707967081218, \"recall\": 0.7327586206896551, \"f1-score\": 0.7332127901216103, \"support\": 116.0}}", + "eval_f1_macro": 48.104679423465, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.32127901216103, + "eval_loss": 0.8219650387763977, + "eval_runtime": 2.1993, + "eval_samples_per_second": 52.743, + "step": 816 + }, + { + "epoch": 52.0, + "learning_rate": 0.0002466666666666666, + "loss": 0.0168, + "step": 832 + }, + { + "epoch": 52.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.40322943340185, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7586206896551724, \"recall\": 0.5945945945945946, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7462973149934357, \"recall\": 0.7949784461881235, \"f1-score\": 0.763627344877345, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7293270789852241, \"recall\": 0.7155172413793104, \"f1-score\": 0.7151701746529333, \"support\": 116.0}}", + "eval_f1_macro": 46.992451992452, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.51701746529334, + "eval_loss": 0.865384042263031, + "eval_runtime": 2.3172, + "eval_samples_per_second": 50.06, + "step": 832 + }, + { + "epoch": 53.0, + "learning_rate": 0.000245, + "loss": 0.0092, + "step": 848 + }, + { + "epoch": 53.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.62124905081899, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7414002843420947, \"recall\": 0.7810895572992347, \"f1-score\": 0.7542136291422947, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.720183698089465, \"recall\": 0.7068965517241379, \"f1-score\": 0.7069253944969183, \"support\": 116.0}}", + "eval_f1_macro": 46.41314640875659, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.69253944969182, + "eval_loss": 0.9035179018974304, + "eval_runtime": 2.2008, + "eval_samples_per_second": 52.708, + "step": 848 + }, + { + "epoch": 54.0, + "learning_rate": 0.0002433333333333333, + "loss": 0.0081, + "step": 864 + }, + { + "epoch": 54.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.59618900858153, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6451612903225806, \"f1-score\": 0.6779661016949152, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6756756756756757, \"f1-score\": 0.6756756756756757, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.751652682902683, \"recall\": 0.7692712874164487, \"f1-score\": 0.756581475304156, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7075786282682834, \"recall\": 0.7068965517241379, \"f1-score\": 0.7044658567078128, \"support\": 116.0}}", + "eval_f1_macro": 46.55886001871728, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.44658567078127, + "eval_loss": 0.8619075417518616, + "eval_runtime": 2.0413, + "eval_samples_per_second": 56.826, + "step": 864 + }, + { + "epoch": 55.0, + "learning_rate": 0.00024166666666666664, + "loss": 0.0174, + "step": 880 + }, + { + "epoch": 55.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.96407889073964, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5625, \"recall\": 0.75, \"f1-score\": 0.6428571428571429, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.8064516129032258, \"f1-score\": 0.7352941176470588, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7952158067998585, \"recall\": 0.7826758209822725, \"f1-score\": 0.7807690879382057, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7501275819498173, \"recall\": 0.7327586206896551, \"f1-score\": 0.7325726293652257, \"support\": 116.0}}", + "eval_f1_macro": 48.04732848850496, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.25726293652257, + "eval_loss": 0.8389707803726196, + "eval_runtime": 2.2977, + "eval_samples_per_second": 50.486, + "step": 880 + }, + { + "epoch": 56.0, + "learning_rate": 0.00023999999999999998, + "loss": 0.0107, + "step": 896 + }, + { + "epoch": 56.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.90321776361796, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.6774193548387096, \"f1-score\": 0.65625, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5555555555555556, \"recall\": 0.5555555555555556, \"f1-score\": 0.5555555555555556, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7431897095959596, \"recall\": 0.766546788724208, \"f1-score\": 0.7518863678671316, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.701200039184953, \"recall\": 0.6982758620689655, \"f1-score\": 0.6968776831039368, \"support\": 116.0}}", + "eval_f1_macro": 46.26993033028503, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.68776831039368, + "eval_loss": 0.8277653455734253, + "eval_runtime": 2.0327, + "eval_samples_per_second": 57.068, + "step": 896 + }, + { + "epoch": 57.0, + "learning_rate": 0.0002383333333333333, + "loss": 0.0078, + "step": 912 + }, + { + "epoch": 57.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.14248466318766, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5, \"recall\": 0.9166666666666666, \"f1-score\": 0.6470588235294118, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8181818181818182, \"recall\": 0.5806451612903226, \"f1-score\": 0.679245283018868, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7708107864357865, \"recall\": 0.7887968613775065, \"f1-score\": 0.7678942896190095, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7512175324675325, \"recall\": 0.7241379310344828, \"f1-score\": 0.7248731923853042, \"support\": 116.0}}", + "eval_f1_macro": 47.25503320732366, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.48731923853042, + "eval_loss": 0.9054758548736572, + "eval_runtime": 2.224, + "eval_samples_per_second": 52.158, + "step": 912 + }, + { + "epoch": 58.0, + "learning_rate": 0.00023666666666666663, + "loss": 0.0117, + "step": 928 + }, + { + "epoch": 58.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.79079973704312, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5263157894736842, \"recall\": 0.8333333333333334, \"f1-score\": 0.6451612903225806, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7532539088905812, \"recall\": 0.7966767170396203, \"f1-score\": 0.7645456154113649, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.728019650860234, \"recall\": 0.7068965517241379, \"f1-score\": 0.7073492765495317, \"support\": 116.0}}", + "eval_f1_macro": 47.048960948391695, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.73492765495317, + "eval_loss": 0.8908520340919495, + "eval_runtime": 2.1575, + "eval_samples_per_second": 53.766, + "step": 928 + }, + { + "epoch": 59.0, + "learning_rate": 0.00023499999999999997, + "loss": 0.0229, + "step": 944 + }, + { + "epoch": 59.0, + "eval_accuracy": 68.10344827586206, + "eval_average_metrics": 62.18145432466472, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5555555555555556, \"recall\": 0.8333333333333334, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6111111111111112, \"recall\": 0.7096774193548387, \"f1-score\": 0.6567164179104478, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.3333333333333333, \"f1-score\": 0.4, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7321428571428571, \"recall\": 0.7464611789208564, \"f1-score\": 0.7288033419213458, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6960864805692392, \"recall\": 0.6810344827586207, \"f1-score\": 0.676694843210058, \"support\": 116.0}}", + "eval_f1_macro": 44.849436425928964, + "eval_f1_micro": 68.10344827586206, + "eval_f1_weighted": 67.6694843210058, + "eval_loss": 0.9175754189491272, + "eval_runtime": 2.3118, + "eval_samples_per_second": 50.177, + "step": 944 + }, + { + "epoch": 60.0, + "learning_rate": 0.0002333333333333333, + "loss": 0.0061, + "step": 960 + }, + { + "epoch": 60.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 63.110820405911554, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5555555555555556, \"recall\": 0.5555555555555556, \"f1-score\": 0.5555555555555556, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7367297537482547, \"recall\": 0.7567840017436792, \"f1-score\": 0.742668905760338, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6946777785454746, \"recall\": 0.6896551724137931, \"f1-score\": 0.6880954524794368, \"support\": 116.0}}", + "eval_f1_macro": 45.70270189294388, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 68.80954524794369, + "eval_loss": 0.8836420178413391, + "eval_runtime": 2.1103, + "eval_samples_per_second": 54.968, + "step": 960 + }, + { + "epoch": 61.0, + "learning_rate": 0.00023166666666666667, + "loss": 0.0147, + "step": 976 + }, + { + "epoch": 61.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.25025679739119, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6153846153846154, \"recall\": 0.6666666666666666, \"f1-score\": 0.64, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6756756756756757, \"f1-score\": 0.6756756756756757, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"request\": {\"precision\": 0.6428571428571429, \"recall\": 1.0, \"f1-score\": 0.782608695652174, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7759658697158698, \"recall\": 0.7906646565920759, \"f1-score\": 0.7771491930240497, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7317981348153763, \"recall\": 0.7241379310344828, \"f1-score\": 0.7234887525811131, \"support\": 116.0}}", + "eval_f1_macro": 47.824565724556905, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.3488752581113, + "eval_loss": 0.8437548875808716, + "eval_runtime": 2.2297, + "eval_samples_per_second": 52.024, + "step": 976 + }, + { + "epoch": 62.0, + "learning_rate": 0.00023, + "loss": 0.0049, + "step": 992 + }, + { + "epoch": 62.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.96754083591335, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6470588235294118, \"recall\": 0.9166666666666666, \"f1-score\": 0.7586206896551724, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6428571428571429, \"recall\": 1.0, \"f1-score\": 0.782608695652174, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7747507605192614, \"recall\": 0.815811779521457, \"f1-score\": 0.78620080953209, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7415756068834654, \"recall\": 0.7327586206896551, \"f1-score\": 0.7293685092682456, \"support\": 116.0}}", + "eval_f1_macro": 48.381588278897844, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 72.93685092682455, + "eval_loss": 0.8428878784179688, + "eval_runtime": 2.0652, + "eval_samples_per_second": 56.169, + "step": 992 + }, + { + "epoch": 63.0, + "learning_rate": 0.0002283333333333333, + "loss": 0.0022, + "step": 1008 + }, + { + "epoch": 63.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.1127357566638, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6486486486486487, \"f1-score\": 0.6956521739130435, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7515151515151515, \"recall\": 0.8006151312602925, \"f1-score\": 0.7680081090407177, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7395898641588295, \"recall\": 0.7241379310344828, \"f1-score\": 0.7236131934032982, \"support\": 116.0}}", + "eval_f1_macro": 47.262037479428784, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.36131934032983, + "eval_loss": 0.8622753024101257, + "eval_runtime": 2.2456, + "eval_samples_per_second": 51.658, + "step": 1008 + }, + { + "epoch": 64.0, + "learning_rate": 0.00022666666666666663, + "loss": 0.002, + "step": 1024 + }, + { + "epoch": 64.0, + "eval_accuracy": 75.86206896551724, + "eval_average_metrics": 69.18501875699872, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.7297297297297297, \"f1-score\": 0.7397260273972601, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7586206896551724, \"macro avg\": {\"precision\": 0.787957702020202, \"recall\": 0.8246391552843165, \"f1-score\": 0.8019321276936633, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7641849529780564, \"recall\": 0.7586206896551724, \"f1-score\": 0.7566626770042725, \"support\": 116.0}}", + "eval_f1_macro": 49.34966939653312, + "eval_f1_micro": 75.86206896551724, + "eval_f1_weighted": 75.66626770042724, + "eval_loss": 0.8323994874954224, + "eval_runtime": 2.1112, + "eval_samples_per_second": 54.945, + "step": 1024 + }, + { + "epoch": 65.0, + "learning_rate": 0.000225, + "loss": 0.0043, + "step": 1040 + }, + { + "epoch": 65.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.35337967908377, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8571428571428571, \"recall\": 0.5806451612903226, \"f1-score\": 0.6923076923076923, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6, \"recall\": 1.0, \"f1-score\": 0.7499999999999999, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7436347926267282, \"recall\": 0.8169500145306596, \"f1-score\": 0.7641936902966314, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7443078288044918, \"recall\": 0.7155172413793104, \"f1-score\": 0.7128276642221876, \"support\": 116.0}}", + "eval_f1_macro": 47.02730401825425, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.28276642221876, + "eval_loss": 1.015073299407959, + "eval_runtime": 2.0888, + "eval_samples_per_second": 55.535, + "step": 1040 + }, + { + "epoch": 66.0, + "learning_rate": 0.00022333333333333333, + "loss": 0.0134, + "step": 1056 + }, + { + "epoch": 66.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.25018013630203, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.55, \"recall\": 0.9166666666666666, \"f1-score\": 0.6874999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7908955627705627, \"recall\": 0.7809018696115471, \"f1-score\": 0.7750247537539606, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7421676742797432, \"recall\": 0.7241379310344828, \"f1-score\": 0.7247930333806785, \"support\": 116.0}}", + "eval_f1_macro": 47.693831000243726, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.47930333806785, + "eval_loss": 0.8981208205223083, + "eval_runtime": 2.2335, + "eval_samples_per_second": 51.937, + "step": 1056 + }, + { + "epoch": 67.0, + "learning_rate": 0.00022166666666666667, + "loss": 0.0019, + "step": 1072 + }, + { + "epoch": 67.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 64.00331673049328, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.5405405405405406, \"f1-score\": 0.6153846153846154, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7590907356532357, \"recall\": 0.7812772449869224, \"f1-score\": 0.7610273954846323, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7102043071870658, \"recall\": 0.6982758620689655, \"f1-score\": 0.6952563940143341, \"support\": 116.0}}", + "eval_f1_macro": 46.8324551067466, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.52563940143341, + "eval_loss": 0.871522068977356, + "eval_runtime": 2.0887, + "eval_samples_per_second": 55.536, + "step": 1072 + }, + { + "epoch": 68.0, + "learning_rate": 0.00021999999999999995, + "loss": 0.0042, + "step": 1088 + }, + { + "epoch": 68.0, + "eval_accuracy": 68.10344827586206, + "eval_average_metrics": 62.42749733192013, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.55, \"recall\": 0.9166666666666666, \"f1-score\": 0.6874999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7307692307692307, \"recall\": 0.5135135135135135, \"f1-score\": 0.6031746031746031, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6810344827586207, \"macro avg\": {\"precision\": 0.7256486568986569, \"recall\": 0.7935798701927734, \"f1-score\": 0.7448186249137336, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.704182455475559, \"recall\": 0.6810344827586207, \"f1-score\": 0.6766810047357272, \"support\": 116.0}}", + "eval_f1_macro": 45.83499230238361, + "eval_f1_micro": 68.10344827586206, + "eval_f1_weighted": 67.66810047357275, + "eval_loss": 0.9840335249900818, + "eval_runtime": 2.1773, + "eval_samples_per_second": 53.276, + "step": 1088 + }, + { + "epoch": 69.0, + "learning_rate": 0.0002183333333333333, + "loss": 0.0116, + "step": 1104 + }, + { + "epoch": 69.0, + "eval_accuracy": 75.0, + "eval_average_metrics": 68.42151422003346, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.7297297297297297, \"f1-score\": 0.7105263157894737, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.7904345654345655, \"recall\": 0.7979814491911266, \"f1-score\": 0.7921106674256406, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.752466929622102, \"recall\": 0.75, \"f1-score\": 0.7494078503855592, \"support\": 116.0}}", + "eval_f1_macro": 48.74527184157789, + "eval_f1_micro": 75.0, + "eval_f1_weighted": 74.94078503855592, + "eval_loss": 0.8145465850830078, + "eval_runtime": 2.2025, + "eval_samples_per_second": 52.668, + "step": 1104 + }, + { + "epoch": 70.0, + "learning_rate": 0.00021666666666666666, + "loss": 0.0124, + "step": 1120 + }, + { + "epoch": 70.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.76443464447134, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5238095238095238, \"recall\": 0.9166666666666666, \"f1-score\": 0.6666666666666667, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7778465821812596, \"recall\": 0.7741451128547903, \"f1-score\": 0.76282674955228, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7291438238254145, \"recall\": 0.7068965517241379, \"f1-score\": 0.7073524364522519, \"support\": 116.0}}", + "eval_f1_macro": 46.94318458783261, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.73524364522518, + "eval_loss": 0.97468101978302, + "eval_runtime": 2.1281, + "eval_samples_per_second": 54.508, + "step": 1120 + }, + { + "epoch": 71.0, + "learning_rate": 0.000215, + "loss": 0.0042, + "step": 1136 + }, + { + "epoch": 71.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.6756069152224, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5675675675675675, \"f1-score\": 0.6363636363636365, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7489322817986611, \"recall\": 0.788127845587523, \"f1-score\": 0.7582478665598871, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7244094713253454, \"recall\": 0.7068965517241379, \"f1-score\": 0.7066171014314591, \"support\": 116.0}}", + "eval_f1_macro": 46.66140717291613, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.66171014314591, + "eval_loss": 0.9142933487892151, + "eval_runtime": 2.1078, + "eval_samples_per_second": 55.033, + "step": 1136 + }, + { + "epoch": 72.0, + "learning_rate": 0.00021333333333333333, + "loss": 0.0025, + "step": 1152 + }, + { + "epoch": 72.0, + "eval_accuracy": 76.72413793103449, + "eval_average_metrics": 70.1131066055635, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7931034482758621, \"recall\": 0.7419354838709677, \"f1-score\": 0.7666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.7297297297297297, \"f1-score\": 0.7297297297297297, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7672413793103449, \"macro avg\": {\"precision\": 0.8292430361395879, \"recall\": 0.8124303739223094, \"f1-score\": 0.8136203167453168, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7809155766944115, \"recall\": 0.7672413793103449, \"f1-score\": 0.7693520799124248, \"support\": 116.0}}", + "eval_f1_macro": 50.06894256894257, + "eval_f1_micro": 76.72413793103448, + "eval_f1_weighted": 76.93520799124248, + "eval_loss": 0.9065346121788025, + "eval_runtime": 2.2637, + "eval_samples_per_second": 51.244, + "step": 1152 + }, + { + "epoch": 73.0, + "learning_rate": 0.00021166666666666667, + "loss": 0.0118, + "step": 1168 + }, + { + "epoch": 73.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.77121142508423, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8076923076923077, \"recall\": 0.6774193548387096, \"f1-score\": 0.7368421052631579, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7902807170899276, \"recall\": 0.808025767703187, \"f1-score\": 0.7879227837145045, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.763951773423189, \"recall\": 0.7413793103448276, \"f1-score\": 0.7432142771047883, \"support\": 116.0}}", + "eval_f1_macro": 48.487555920892575, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.32142771047884, + "eval_loss": 1.009997844696045, + "eval_runtime": 2.2483, + "eval_samples_per_second": 51.595, + "step": 1168 + }, + { + "epoch": 74.0, + "learning_rate": 0.00020999999999999998, + "loss": 0.0026, + "step": 1184 + }, + { + "epoch": 74.0, + "eval_accuracy": 76.72413793103449, + "eval_average_metrics": 70.049578589584, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.7741935483870968, \"f1-score\": 0.7619047619047619, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7672413793103449, \"macro avg\": {\"precision\": 0.8173160173160172, \"recall\": 0.8235947641189576, \"f1-score\": 0.8147149548078341, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7732907896701001, \"recall\": 0.7672413793103449, \"f1-score\": 0.7661373358501574, \"support\": 116.0}}", + "eval_f1_macro": 50.13630491125133, + "eval_f1_micro": 76.72413793103448, + "eval_f1_weighted": 76.61373358501574, + "eval_loss": 0.8809800148010254, + "eval_runtime": 2.1396, + "eval_samples_per_second": 54.215, + "step": 1184 + }, + { + "epoch": 75.0, + "learning_rate": 0.00020833333333333332, + "loss": 0.008, + "step": 1200 + }, + { + "epoch": 75.0, + "eval_accuracy": 76.72413793103449, + "eval_average_metrics": 70.02564938184499, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8518518518518519, \"recall\": 0.7419354838709677, \"f1-score\": 0.7931034482758621, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.6428571428571429, \"recall\": 1.0, \"f1-score\": 0.782608695652174, \"support\": 9.0}, \"accuracy\": 0.7672413793103449, \"macro avg\": {\"precision\": 0.8030321530321531, \"recall\": 0.8334513949433304, \"f1-score\": 0.8119859478155659, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7759448724965966, \"recall\": 0.7672413793103449, \"f1-score\": 0.7668595564589151, \"support\": 116.0}}", + "eval_f1_macro": 49.96836601941944, + "eval_f1_micro": 76.72413793103448, + "eval_f1_weighted": 76.68595564589151, + "eval_loss": 0.9097422957420349, + "eval_runtime": 2.1959, + "eval_samples_per_second": 52.826, + "step": 1200 + }, + { + "epoch": 76.0, + "learning_rate": 0.00020666666666666666, + "loss": 0.0013, + "step": 1216 + }, + { + "epoch": 76.0, + "eval_accuracy": 75.0, + "eval_average_metrics": 68.43649890784998, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8148148148148148, \"recall\": 0.7096774193548387, \"f1-score\": 0.7586206896551724, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.7027027027027027, \"f1-score\": 0.7222222222222223, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.75, \"macro avg\": {\"precision\": 0.7949074074074074, \"recall\": 0.8015475152571927, \"f1-score\": 0.7897184592227696, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7666849115124977, \"recall\": 0.75, \"f1-score\": 0.751479366023064, \"support\": 116.0}}", + "eval_f1_macro": 48.59805902909351, + "eval_f1_micro": 75.0, + "eval_f1_weighted": 75.1479366023064, + "eval_loss": 0.9198105931282043, + "eval_runtime": 2.12, + "eval_samples_per_second": 54.717, + "step": 1216 + }, + { + "epoch": 77.0, + "learning_rate": 0.000205, + "loss": 0.0034, + "step": 1232 + }, + { + "epoch": 77.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.9730843369445, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.7419354838709677, \"f1-score\": 0.7419354838709677, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7723275828114539, \"recall\": 0.802389082631018, \"f1-score\": 0.781404294366818, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.742217390507157, \"recall\": 0.7327586206896551, \"f1-score\": 0.7325419509496588, \"support\": 116.0}}", + "eval_f1_macro": 48.086418114881106, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.25419509496588, + "eval_loss": 0.9556060433387756, + "eval_runtime": 2.2754, + "eval_samples_per_second": 50.979, + "step": 1232 + }, + { + "epoch": 78.0, + "learning_rate": 0.00020333333333333333, + "loss": 0.0014, + "step": 1248 + }, + { + "epoch": 78.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.08748867920372, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.6216216216216216, \"f1-score\": 0.696969696969697, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.743769341290893, \"recall\": 0.8047412331686525, \"f1-score\": 0.7655129972574192, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7420494273199386, \"recall\": 0.7241379310344828, \"f1-score\": 0.7241387637100023, \"support\": 116.0}}", + "eval_f1_macro": 47.1084921389181, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.41387637100023, + "eval_loss": 1.0091474056243896, + "eval_runtime": 2.0996, + "eval_samples_per_second": 55.249, + "step": 1248 + }, + { + "epoch": 79.0, + "learning_rate": 0.00020166666666666667, + "loss": 0.003, + "step": 1264 + }, + { + "epoch": 79.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.80793026911971, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.7741935483870968, \"f1-score\": 0.7384615384615385, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.6216216216216216, \"f1-score\": 0.7076923076923075, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7594654528478058, \"recall\": 0.7925324518066452, \"f1-score\": 0.7713542670779513, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7447370327441322, \"recall\": 0.7327586206896551, \"f1-score\": 0.7321204204144313, \"support\": 116.0}}", + "eval_f1_macro": 47.46795489710468, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.21204204144313, + "eval_loss": 1.009720802307129, + "eval_runtime": 2.21, + "eval_samples_per_second": 52.488, + "step": 1264 + }, + { + "epoch": 80.0, + "learning_rate": 0.00019999999999999998, + "loss": 0.001, + "step": 1280 + }, + { + "epoch": 80.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.2278718761285, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7611828695498051, \"recall\": 0.8047412331686525, \"f1-score\": 0.7747153145479717, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.738864900085701, \"recall\": 0.7241379310344828, \"f1-score\": 0.7240911271004996, \"support\": 116.0}}", + "eval_f1_macro": 47.67478858756749, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.40911271004997, + "eval_loss": 1.0751197338104248, + "eval_runtime": 2.1451, + "eval_samples_per_second": 54.077, + "step": 1280 + }, + { + "epoch": 81.0, + "learning_rate": 0.00019833333333333332, + "loss": 0.0034, + "step": 1296 + }, + { + "epoch": 81.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.67391316856995, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.7419354838709677, \"f1-score\": 0.7076923076923077, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7777777777777778, \"recall\": 0.5675675675675675, \"f1-score\": 0.65625, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7519580909286792, \"recall\": 0.7817434369853724, \"f1-score\": 0.7597936922361264, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7219307793851404, \"recall\": 0.7068965517241379, \"f1-score\": 0.7055980742261369, \"support\": 116.0}}", + "eval_f1_macro": 46.756534906838546, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.55980742261369, + "eval_loss": 1.0636241436004639, + "eval_runtime": 2.1338, + "eval_samples_per_second": 54.363, + "step": 1296 + }, + { + "epoch": 82.0, + "learning_rate": 0.00019666666666666666, + "loss": 0.0042, + "step": 1312 + }, + { + "epoch": 82.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.11615353179951, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7566690066690066, \"recall\": 0.7929229632858665, \"f1-score\": 0.7692698861816509, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7331784594715629, \"recall\": 0.7241379310344828, \"f1-score\": 0.7229734261681524, \"support\": 116.0}}", + "eval_f1_macro": 47.33968530348621, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.29734261681524, + "eval_loss": 1.0073676109313965, + "eval_runtime": 2.1573, + "eval_samples_per_second": 53.77, + "step": 1312 + }, + { + "epoch": 83.0, + "learning_rate": 0.000195, + "loss": 0.0031, + "step": 1328 + }, + { + "epoch": 83.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.7525479876361, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7668492461839236, \"recall\": 0.771232926474862, \"f1-score\": 0.7617192274637128, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7197790506675379, \"recall\": 0.7068965517241379, \"f1-score\": 0.7075585222333448, \"support\": 116.0}}", + "eval_f1_macro": 46.87502938238233, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.75585222333449, + "eval_loss": 1.0074830055236816, + "eval_runtime": 2.3548, + "eval_samples_per_second": 49.262, + "step": 1328 + }, + { + "epoch": 84.0, + "learning_rate": 0.00019333333333333333, + "loss": 0.0023, + "step": 1344 + }, + { + "epoch": 84.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.96007923927792, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5, \"recall\": 0.5555555555555556, \"f1-score\": 0.5263157894736842, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7552319004524887, \"recall\": 0.7630745665019858, \"f1-score\": 0.7510372788328765, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7128695974410986, \"recall\": 0.6982758620689655, \"f1-score\": 0.6996746584591074, \"support\": 116.0}}", + "eval_f1_macro": 46.21767869740778, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.96746584591075, + "eval_loss": 0.9842262864112854, + "eval_runtime": 2.2804, + "eval_samples_per_second": 50.867, + "step": 1344 + }, + { + "epoch": 85.0, + "learning_rate": 0.00019166666666666665, + "loss": 0.0013, + "step": 1360 + }, + { + "epoch": 85.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.34671058116388, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.875, \"f1-score\": 0.7368421052631579, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.6216216216216216, \"f1-score\": 0.6865671641791045, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7349621212121212, \"recall\": 0.8123017170396203, \"f1-score\": 0.7641747663104694, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7281687565308255, \"recall\": 0.7155172413793104, \"f1-score\": 0.712572545835338, \"support\": 116.0}}", + "eval_f1_macro": 47.02613946525966, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.25725458353381, + "eval_loss": 0.9387104511260986, + "eval_runtime": 2.2626, + "eval_samples_per_second": 51.269, + "step": 1360 + }, + { + "epoch": 86.0, + "learning_rate": 0.00018999999999999998, + "loss": 0.0014, + "step": 1376 + }, + { + "epoch": 86.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0203074746172, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7808308929632459, \"recall\": 0.7982629807226582, \"f1-score\": 0.7830183865275773, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7437090039777667, \"recall\": 0.7327586206896551, \"f1-score\": 0.7334375889730225, \"support\": 116.0}}", + "eval_f1_macro": 48.18574686323552, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.34375889730225, + "eval_loss": 0.9032027125358582, + "eval_runtime": 2.3121, + "eval_samples_per_second": 50.17, + "step": 1376 + }, + { + "epoch": 87.0, + "learning_rate": 0.00018833333333333332, + "loss": 0.0005, + "step": 1392 + }, + { + "epoch": 87.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.12735809570236, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8076923076923077, \"recall\": 0.6774193548387096, \"f1-score\": 0.7368421052631579, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7869123931623933, \"recall\": 0.8081196115470308, \"f1-score\": 0.7890888351414667, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7480153467222432, \"recall\": 0.7327586206896551, \"f1-score\": 0.7339839531309585, \"support\": 116.0}}", + "eval_f1_macro": 48.559312931782564, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.39839531309586, + "eval_loss": 0.9623767733573914, + "eval_runtime": 2.1627, + "eval_samples_per_second": 53.637, + "step": 1392 + }, + { + "epoch": 88.0, + "learning_rate": 0.00018666666666666666, + "loss": 0.0008, + "step": 1408 + }, + { + "epoch": 88.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.55394117974937, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.875, \"f1-score\": 0.7777777777777777, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.8260869565217391, \"recall\": 0.5135135135135135, \"f1-score\": 0.6333333333333333, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7372230879024357, \"recall\": 0.8010283468952824, \"f1-score\": 0.7575385551948053, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7303347452148053, \"recall\": 0.7068965517241379, \"f1-score\": 0.7021869713141263, \"support\": 116.0}}", + "eval_f1_macro": 46.61775724275724, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.21869713141264, + "eval_loss": 1.013329029083252, + "eval_runtime": 2.0721, + "eval_samples_per_second": 55.981, + "step": 1408 + }, + { + "epoch": 89.0, + "learning_rate": 0.000185, + "loss": 0.0006, + "step": 1424 + }, + { + "epoch": 89.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.65086863670423, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6756756756756757, \"f1-score\": 0.7142857142857142, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7740124458874459, \"recall\": 0.8016413591010365, \"f1-score\": 0.7816112509033746, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7531861098671444, \"recall\": 0.7413793103448276, \"f1-score\": 0.7422845857610528, \"support\": 116.0}}", + "eval_f1_macro": 48.09915390174613, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.22845857610528, + "eval_loss": 0.9453557729721069, + "eval_runtime": 2.079, + "eval_samples_per_second": 55.796, + "step": 1424 + }, + { + "epoch": 90.0, + "learning_rate": 0.00018333333333333334, + "loss": 0.0012, + "step": 1440 + }, + { + "epoch": 90.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.21906865002549, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7661008313581843, \"recall\": 0.794230722658142, \"f1-score\": 0.77336173049969, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7355240615255829, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245719728553219, \"support\": 116.0}}", + "eval_f1_macro": 47.591491107673235, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.45719728553219, + "eval_loss": 0.988357424736023, + "eval_runtime": 2.409, + "eval_samples_per_second": 48.153, + "step": 1440 + }, + { + "epoch": 91.0, + "learning_rate": 0.00018166666666666665, + "loss": 0.0058, + "step": 1456 + }, + { + "epoch": 91.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.72942944952882, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6875, \"recall\": 0.5945945945945946, \"f1-score\": 0.6376811594202898, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7526423229548229, \"recall\": 0.7874739659013852, \"f1-score\": 0.7616967013055733, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7204242716096164, \"recall\": 0.7068965517241379, \"f1-score\": 0.7066476429602169, \"support\": 116.0}}", + "eval_f1_macro": 46.873643157266045, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.6647642960217, + "eval_loss": 1.003797173500061, + "eval_runtime": 2.3032, + "eval_samples_per_second": 50.364, + "step": 1456 + }, + { + "epoch": 92.0, + "learning_rate": 0.00017999999999999998, + "loss": 0.0037, + "step": 1472 + }, + { + "epoch": 92.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.49007788853585, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8076923076923077, \"recall\": 0.6774193548387096, \"f1-score\": 0.7368421052631579, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7602112960568843, \"recall\": 0.7908523442797636, \"f1-score\": 0.7678673719564575, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7292110146445846, \"recall\": 0.7155172413793104, \"f1-score\": 0.7160348654249928, \"support\": 116.0}}", + "eval_f1_macro": 47.253376735782005, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.6034865424993, + "eval_loss": 0.9861951470375061, + "eval_runtime": 2.189, + "eval_samples_per_second": 52.993, + "step": 1472 + }, + { + "epoch": 93.0, + "learning_rate": 0.00017833333333333332, + "loss": 0.0004, + "step": 1488 + }, + { + "epoch": 93.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0835534197603, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8333333333333334, \"recall\": 0.8333333333333334, \"f1-score\": 0.8333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.7096774193548387, \"f1-score\": 0.7333333333333333, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7792091768384872, \"recall\": 0.8087734912331687, \"f1-score\": 0.7883297258297258, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7409545220930478, \"recall\": 0.7327586206896551, \"f1-score\": 0.7326989102851171, \"support\": 116.0}}", + "eval_f1_macro": 48.512598512598515, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.26989102851171, + "eval_loss": 0.9672034978866577, + "eval_runtime": 2.3052, + "eval_samples_per_second": 50.322, + "step": 1488 + }, + { + "epoch": 94.0, + "learning_rate": 0.00017666666666666666, + "loss": 0.0047, + "step": 1504 + }, + { + "epoch": 94.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.61081018532731, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8333333333333334, \"recall\": 0.8333333333333334, \"f1-score\": 0.8333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.5675675675675675, \"f1-score\": 0.626865671641791, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5714285714285714, \"recall\": 0.4444444444444444, \"f1-score\": 0.5, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7643533643533644, \"recall\": 0.7684145839387775, \"f1-score\": 0.7602803717597397, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7116083590221521, \"recall\": 0.7068965517241379, \"f1-score\": 0.7027744598049765, \"support\": 116.0}}", + "eval_f1_macro": 46.78648441598398, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.27744598049766, + "eval_loss": 0.986122190952301, + "eval_runtime": 2.1905, + "eval_samples_per_second": 52.957, + "step": 1504 + }, + { + "epoch": 95.0, + "learning_rate": 0.000175, + "loss": 0.0018, + "step": 1520 + }, + { + "epoch": 95.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.83607096081741, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8333333333333334, \"recall\": 0.8333333333333334, \"f1-score\": 0.8333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6486486486486487, \"recall\": 0.7741935483870968, \"f1-score\": 0.7058823529411764, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6774193548387096, \"recall\": 0.5675675675675675, \"f1-score\": 0.6176470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5714285714285714, \"recall\": 0.4444444444444444, \"f1-score\": 0.5, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7597946725970919, \"recall\": 0.7545256950498886, \"f1-score\": 0.7509581105169341, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7033283256517405, \"recall\": 0.6982758620689655, \"f1-score\": 0.6947630462843445, \"support\": 116.0}}", + "eval_f1_macro": 46.21280680104209, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.47630462843445, + "eval_loss": 0.9718347787857056, + "eval_runtime": 2.3653, + "eval_samples_per_second": 49.042, + "step": 1520 + }, + { + "epoch": 96.0, + "learning_rate": 0.0001733333333333333, + "loss": 0.0007, + "step": 1536 + }, + { + "epoch": 96.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.76175141658041, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8333333333333334, \"recall\": 0.8333333333333334, \"f1-score\": 0.8333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7666666666666667, \"recall\": 0.7419354838709677, \"f1-score\": 0.7540983606557377, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7860031512605042, \"recall\": 0.8022952387871742, \"f1-score\": 0.7894486090441719, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7488487636433885, \"recall\": 0.7413793103448276, \"f1-score\": 0.7418969073309942, \"support\": 116.0}}", + "eval_f1_macro": 48.581452864256725, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.18969073309943, + "eval_loss": 0.9660546779632568, + "eval_runtime": 2.1618, + "eval_samples_per_second": 53.659, + "step": 1536 + }, + { + "epoch": 97.0, + "learning_rate": 0.00017166666666666665, + "loss": 0.0005, + "step": 1552 + }, + { + "epoch": 97.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.37793808045248, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.6486486486486487, \"f1-score\": 0.6575342465753425, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7854662698412698, \"recall\": 0.794230722658142, \"f1-score\": 0.7825183305399483, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7368466064586755, \"recall\": 0.7241379310344828, \"f1-score\": 0.7252919192783961, \"support\": 116.0}}", + "eval_f1_macro": 48.15497418707374, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.5291919278396, + "eval_loss": 0.9771972894668579, + "eval_runtime": 2.1794, + "eval_samples_per_second": 53.226, + "step": 1552 + }, + { + "epoch": 98.0, + "learning_rate": 0.00016999999999999999, + "loss": 0.0034, + "step": 1568 + }, + { + "epoch": 98.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.7984190383411, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5675675675675675, \"f1-score\": 0.6363636363636365, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7658504060216855, \"recall\": 0.7945122541896734, \"f1-score\": 0.7665646406681366, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7297752879981898, \"recall\": 0.7068965517241379, \"f1-score\": 0.7064115715203608, \"support\": 116.0}}", + "eval_f1_macro": 47.17320865650072, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.6411571520361, + "eval_loss": 1.0846428871154785, + "eval_runtime": 2.1972, + "eval_samples_per_second": 52.795, + "step": 1568 + }, + { + "epoch": 99.0, + "learning_rate": 0.00016833333333333332, + "loss": 0.0011, + "step": 1584 + }, + { + "epoch": 99.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.07366957347286, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7878336940836941, \"recall\": 0.7976091010365204, \"f1-score\": 0.7850258781673589, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7475011817684232, \"recall\": 0.7327586206896551, \"f1-score\": 0.7343366934566137, \"support\": 116.0}}", + "eval_f1_macro": 48.309284810299005, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.43366934566137, + "eval_loss": 0.9998502135276794, + "eval_runtime": 2.0893, + "eval_samples_per_second": 55.521, + "step": 1584 + }, + { + "epoch": 100.0, + "learning_rate": 0.00016666666666666666, + "loss": 0.0058, + "step": 1600 + }, + { + "epoch": 100.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.84299596892572, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.7567567567567568, \"f1-score\": 0.7272727272727273, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.8226190476190476, \"recall\": 0.7799664583938777, \"f1-score\": 0.7944732723097693, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7519704433497537, \"recall\": 0.7413793103448276, \"f1-score\": 0.7420545889536696, \"support\": 116.0}}", + "eval_f1_macro": 48.89066291137042, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.20545889536697, + "eval_loss": 0.9746664762496948, + "eval_runtime": 2.2781, + "eval_samples_per_second": 50.92, + "step": 1600 + }, + { + "epoch": 101.0, + "learning_rate": 0.000165, + "loss": 0.0019, + "step": 1616 + }, + { + "epoch": 101.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.95791619492573, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7857142857142857, \"recall\": 0.9166666666666666, \"f1-score\": 0.8461538461538461, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5714285714285714, \"recall\": 0.4444444444444444, \"f1-score\": 0.5, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.8001779878618114, \"recall\": 0.7737697374794148, \"f1-score\": 0.7822729027832086, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7364437441643324, \"recall\": 0.7327586206896551, \"f1-score\": 0.7314006970126677, \"support\": 116.0}}", + "eval_f1_macro": 48.13987094050513, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.14006970126677, + "eval_loss": 1.0031696557998657, + "eval_runtime": 2.174, + "eval_samples_per_second": 53.358, + "step": 1616 + }, + { + "epoch": 102.0, + "learning_rate": 0.0001633333333333333, + "loss": 0.0003, + "step": 1632 + }, + { + "epoch": 102.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.6906117356637, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.6486486486486487, \"f1-score\": 0.6575342465753425, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7468434343434344, \"recall\": 0.7861662065291097, \"f1-score\": 0.7613133536934906, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7153539707419017, \"recall\": 0.7068965517241379, \"f1-score\": 0.7053308406284317, \"support\": 116.0}}", + "eval_f1_macro": 46.85005253498404, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.53308406284316, + "eval_loss": 1.0247207880020142, + "eval_runtime": 2.125, + "eval_samples_per_second": 54.588, + "step": 1632 + }, + { + "epoch": 103.0, + "learning_rate": 0.00016166666666666665, + "loss": 0.0076, + "step": 1648 + }, + { + "epoch": 103.0, + "eval_accuracy": 67.24137931034483, + "eval_average_metrics": 61.543610356616945, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5789473684210527, \"recall\": 0.9166666666666666, \"f1-score\": 0.7096774193548387, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.6774193548387096, \"f1-score\": 0.65625, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.76, \"recall\": 0.5135135135135135, \"f1-score\": 0.6129032258064517, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5555555555555556, \"recall\": 0.5555555555555556, \"f1-score\": 0.5555555555555556, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6724137931034483, \"macro avg\": {\"precision\": 0.7214063969656075, \"recall\": 0.7599777196551389, \"f1-score\": 0.7293985424692849, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6914491610929905, \"recall\": 0.6724137931034483, \"f1-score\": 0.6680561865382214, \"support\": 116.0}}", + "eval_f1_macro": 44.88606415195599, + "eval_f1_micro": 67.24137931034483, + "eval_f1_weighted": 66.80561865382214, + "eval_loss": 1.1534621715545654, + "eval_runtime": 2.1588, + "eval_samples_per_second": 53.734, + "step": 1648 + }, + { + "epoch": 104.0, + "learning_rate": 0.00015999999999999999, + "loss": 0.001, + "step": 1664 + }, + { + "epoch": 104.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.83670185178806, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6571428571428571, \"recall\": 0.6216216216216216, \"f1-score\": 0.6388888888888888, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7589486901986902, \"recall\": 0.7966767170396203, \"f1-score\": 0.7700306637806638, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7196450819295647, \"recall\": 0.7068965517241379, \"f1-score\": 0.7058099467582226, \"support\": 116.0}}", + "eval_f1_macro": 47.386502386502386, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.58099467582227, + "eval_loss": 1.0140974521636963, + "eval_runtime": 2.0908, + "eval_samples_per_second": 55.482, + "step": 1664 + }, + { + "epoch": 105.0, + "learning_rate": 0.00015833333333333332, + "loss": 0.0019, + "step": 1680 + }, + { + "epoch": 105.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.47830212383965, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7583858311229001, \"recall\": 0.7908523442797636, \"f1-score\": 0.7697431245225362, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7207556736763039, \"recall\": 0.7155172413793104, \"f1-score\": 0.7144095255657122, \"support\": 116.0}}", + "eval_f1_macro": 47.36880766292531, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.44095255657122, + "eval_loss": 0.9925776124000549, + "eval_runtime": 2.1371, + "eval_samples_per_second": 54.278, + "step": 1680 + }, + { + "epoch": 106.0, + "learning_rate": 0.00015666666666666666, + "loss": 0.0006, + "step": 1696 + }, + { + "epoch": 106.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.27571069810807, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6451612903225806, \"f1-score\": 0.6779661016949152, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7708246614496614, \"recall\": 0.7935768429720043, \"f1-score\": 0.7795533775386171, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7259982594034318, \"recall\": 0.7241379310344828, \"f1-score\": 0.7230274104469775, \"support\": 116.0}}", + "eval_f1_macro": 47.97251554083797, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.30274104469775, + "eval_loss": 0.9963064789772034, + "eval_runtime": 2.1179, + "eval_samples_per_second": 54.77, + "step": 1696 + }, + { + "epoch": 107.0, + "learning_rate": 0.000155, + "loss": 0.0024, + "step": 1712 + }, + { + "epoch": 107.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.7377253700138, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.631578947368421, \"recall\": 0.7741935483870968, \"f1-score\": 0.6956521739130435, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7779858299595142, \"recall\": 0.7613762956504893, \"f1-score\": 0.7628976145941829, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7199997673228162, \"recall\": 0.7068965517241379, \"f1-score\": 0.7062404562173948, \"support\": 116.0}}", + "eval_f1_macro": 46.94754551348818, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.62404562173948, + "eval_loss": 1.0944117307662964, + "eval_runtime": 2.2195, + "eval_samples_per_second": 52.264, + "step": 1712 + }, + { + "epoch": 108.0, + "learning_rate": 0.0001533333333333333, + "loss": 0.0002, + "step": 1728 + }, + { + "epoch": 108.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.35003338809494, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.7948846023442797, \"f1-score\": 0.7827215130132199, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.733823145245559, \"recall\": 0.7241379310344828, \"f1-score\": 0.7240506962159275, \"support\": 116.0}}", + "eval_f1_macro": 48.16747772389046, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.40506962159276, + "eval_loss": 1.0319750308990479, + "eval_runtime": 2.2528, + "eval_samples_per_second": 51.491, + "step": 1728 + }, + { + "epoch": 109.0, + "learning_rate": 0.00015166666666666665, + "loss": 0.0013, + "step": 1744 + }, + { + "epoch": 109.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.66781775563415, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6875, \"recall\": 0.5945945945945946, \"f1-score\": 0.6376811594202898, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7836538461538461, \"recall\": 0.7915062239659014, \"f1-score\": 0.7806671944139975, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7250497347480106, \"recall\": 0.7155172413793104, \"f1-score\": 0.7152676462889004, \"support\": 116.0}}", + "eval_f1_macro": 48.04105811778446, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.52676462889004, + "eval_loss": 1.0217504501342773, + "eval_runtime": 2.24, + "eval_samples_per_second": 51.786, + "step": 1744 + }, + { + "epoch": 110.0, + "learning_rate": 0.00015, + "loss": 0.0011, + "step": 1760 + }, + { + "epoch": 110.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.81005930236041, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7857142857142857, \"recall\": 0.7096774193548387, \"f1-score\": 0.7457627118644068, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7940340909090908, \"recall\": 0.8016413591010365, \"f1-score\": 0.7910358540779776, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7539324899238693, \"recall\": 0.7413793103448276, \"f1-score\": 0.7428524565875445, \"support\": 116.0}}", + "eval_f1_macro": 48.6791294817217, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.28524565875443, + "eval_loss": 1.0107359886169434, + "eval_runtime": 2.059, + "eval_samples_per_second": 56.337, + "step": 1760 + }, + { + "epoch": 111.0, + "learning_rate": 0.00014833333333333332, + "loss": 0.0006, + "step": 1776 + }, + { + "epoch": 111.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.72662641030664, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5945945945945946, \"f1-score\": 0.6470588235294118, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7548846482818258, \"recall\": 0.7810895572992347, \"f1-score\": 0.761943627912298, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7159673092480378, \"recall\": 0.7068965517241379, \"f1-score\": 0.7063835665564212, \"support\": 116.0}}", + "eval_f1_macro": 46.888838640756795, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.63835665564213, + "eval_loss": 1.0691028833389282, + "eval_runtime": 2.2902, + "eval_samples_per_second": 50.652, + "step": 1776 + }, + { + "epoch": 112.0, + "learning_rate": 0.00014666666666666664, + "loss": 0.0003, + "step": 1792 + }, + { + "epoch": 112.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.27387783514109, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7770055465367965, \"recall\": 0.7885001937421292, \"f1-score\": 0.7775157753318719, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7327597869084939, \"recall\": 0.7241379310344828, \"f1-score\": 0.7242080049786032, \"support\": 116.0}}", + "eval_f1_macro": 47.8471246358075, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.42080049786031, + "eval_loss": 1.070449709892273, + "eval_runtime": 2.1308, + "eval_samples_per_second": 54.439, + "step": 1792 + }, + { + "epoch": 113.0, + "learning_rate": 0.000145, + "loss": 0.0013, + "step": 1808 + }, + { + "epoch": 113.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.94971808239175, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7709749740999741, \"recall\": 0.7969552213503825, \"f1-score\": 0.7802725022460566, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7378138703569739, \"recall\": 0.7327586206896551, \"f1-score\": 0.7323037882264791, \"support\": 116.0}}", + "eval_f1_macro": 48.0167693689881, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.2303788226479, + "eval_loss": 1.0506659746170044, + "eval_runtime": 2.1797, + "eval_samples_per_second": 53.219, + "step": 1808 + }, + { + "epoch": 114.0, + "learning_rate": 0.00014333333333333334, + "loss": 0.0014, + "step": 1824 + }, + { + "epoch": 114.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.76925999714823, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7625518476621418, \"recall\": 0.788127845587523, \"f1-score\": 0.7626786408151753, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7310497607531076, \"recall\": 0.7068965517241379, \"f1-score\": 0.7076365943975451, \"support\": 116.0}}", + "eval_f1_macro": 46.934070204010794, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.7636594397545, + "eval_loss": 1.1001496315002441, + "eval_runtime": 2.131, + "eval_samples_per_second": 54.435, + "step": 1824 + }, + { + "epoch": 115.0, + "learning_rate": 0.00014166666666666665, + "loss": 0.0007, + "step": 1840 + }, + { + "epoch": 115.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.70115661864702, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7503036791639733, \"recall\": 0.7874739659013852, \"f1-score\": 0.7594631003510002, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7238169073995647, \"recall\": 0.7068965517241379, \"f1-score\": 0.7068912533892974, \"support\": 116.0}}", + "eval_f1_macro": 46.736190790830776, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.68912533892974, + "eval_loss": 1.0808982849121094, + "eval_runtime": 2.193, + "eval_samples_per_second": 52.895, + "step": 1840 + }, + { + "epoch": 116.0, + "learning_rate": 0.00014, + "loss": 0.0002, + "step": 1856 + }, + { + "epoch": 116.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.88409380563921, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.727754884004884, \"recall\": 0.7932983386612419, \"f1-score\": 0.7513110312428937, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7116505620815966, \"recall\": 0.6982758620689655, \"f1-score\": 0.6964667780920103, \"support\": 116.0}}", + "eval_f1_macro": 46.23452499956269, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.64667780920102, + "eval_loss": 1.1061365604400635, + "eval_runtime": 2.1218, + "eval_samples_per_second": 54.67, + "step": 1856 + }, + { + "epoch": 117.0, + "learning_rate": 0.00013833333333333333, + "loss": 0.0058, + "step": 1872 + }, + { + "epoch": 117.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 63.282710359282376, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7441620879120879, \"recall\": 0.7794094497723529, \"f1-score\": 0.7533740769034887, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7012165277251484, \"recall\": 0.6896551724137931, \"f1-score\": 0.6883832529877154, \"support\": 116.0}}", + "eval_f1_macro": 46.3614816555993, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 68.83832529877154, + "eval_loss": 1.0625555515289307, + "eval_runtime": 2.1789, + "eval_samples_per_second": 53.238, + "step": 1872 + }, + { + "epoch": 118.0, + "learning_rate": 0.00013666666666666666, + "loss": 0.0003, + "step": 1888 + }, + { + "epoch": 118.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 63.282710359282376, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.625, \"recall\": 0.8333333333333334, \"f1-score\": 0.7142857142857143, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7307692307692307, \"recall\": 0.6129032258064516, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.7441620879120879, \"recall\": 0.7794094497723529, \"f1-score\": 0.7533740769034887, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7012165277251484, \"recall\": 0.6896551724137931, \"f1-score\": 0.6883832529877154, \"support\": 116.0}}", + "eval_f1_macro": 46.3614816555993, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 68.83832529877154, + "eval_loss": 1.0653810501098633, + "eval_runtime": 2.1294, + "eval_samples_per_second": 54.477, + "step": 1888 + }, + { + "epoch": 119.0, + "learning_rate": 0.000135, + "loss": 0.0026, + "step": 1904 + }, + { + "epoch": 119.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.66956317091271, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7840403883750657, \"recall\": 0.7915062239659014, \"f1-score\": 0.7807742305133197, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.72446698416526, \"recall\": 0.7155172413793104, \"f1-score\": 0.7152715945312295, \"support\": 116.0}}", + "eval_f1_macro": 48.04764495466583, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.52715945312295, + "eval_loss": 1.0476528406143188, + "eval_runtime": 2.0205, + "eval_samples_per_second": 57.412, + "step": 1904 + }, + { + "epoch": 120.0, + "learning_rate": 0.0001333333333333333, + "loss": 0.003, + "step": 1920 + }, + { + "epoch": 120.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.70006170794646, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7461895102868734, \"recall\": 0.7979844764118957, \"f1-score\": 0.7633107137871202, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7203813986950309, \"recall\": 0.7068965517241379, \"f1-score\": 0.7044796948467391, \"support\": 116.0}}", + "eval_f1_macro": 46.97296700228433, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.4479694846739, + "eval_loss": 1.0797429084777832, + "eval_runtime": 2.0702, + "eval_samples_per_second": 56.032, + "step": 1920 + }, + { + "epoch": 121.0, + "learning_rate": 0.00013166666666666665, + "loss": 0.0049, + "step": 1936 + }, + { + "epoch": 121.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.70006170794646, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5882352941176471, \"recall\": 0.8333333333333334, \"f1-score\": 0.6896551724137931, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.5675675675675675, \"f1-score\": 0.6461538461538462, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7461895102868734, \"recall\": 0.7979844764118957, \"f1-score\": 0.7633107137871202, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7203813986950309, \"recall\": 0.7068965517241379, \"f1-score\": 0.7044796948467391, \"support\": 116.0}}", + "eval_f1_macro": 46.97296700228433, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.4479694846739, + "eval_loss": 1.094364881515503, + "eval_runtime": 2.0626, + "eval_samples_per_second": 56.239, + "step": 1936 + }, + { + "epoch": 122.0, + "learning_rate": 0.00013, + "loss": 0.0013, + "step": 1952 + }, + { + "epoch": 122.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.88829920079921, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6111111111111112, \"recall\": 0.9166666666666666, \"f1-score\": 0.7333333333333334, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5675675675675675, \"f1-score\": 0.6363636363636365, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.734797821909891, \"recall\": 0.8003366269495301, \"f1-score\": 0.755275974025974, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7144611700230011, \"recall\": 0.6982758620689655, \"f1-score\": 0.6941950291088222, \"support\": 116.0}}", + "eval_f1_macro": 46.47852147852148, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.41950291088223, + "eval_loss": 1.1192301511764526, + "eval_runtime": 2.2276, + "eval_samples_per_second": 52.075, + "step": 1952 + }, + { + "epoch": 123.0, + "learning_rate": 0.00012833333333333333, + "loss": 0.0002, + "step": 1968 + }, + { + "epoch": 123.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 64.06779904745557, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.5, \"recall\": 0.9166666666666666, \"f1-score\": 0.6470588235294118, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8571428571428571, \"recall\": 0.5806451612903226, \"f1-score\": 0.6923076923076923, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7333333333333333, \"recall\": 0.5945945945945946, \"f1-score\": 0.6567164179104478, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7556547619047619, \"recall\": 0.7996827472633924, \"f1-score\": 0.7572062368285017, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7439449917898194, \"recall\": 0.6982758620689655, \"f1-score\": 0.7001871689427523, \"support\": 116.0}}", + "eval_f1_macro": 46.59730688175395, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 70.01871689427523, + "eval_loss": 1.1883636713027954, + "eval_runtime": 2.2785, + "eval_samples_per_second": 50.911, + "step": 1968 + }, + { + "epoch": 124.0, + "learning_rate": 0.00012666666666666666, + "loss": 0.001, + "step": 1984 + }, + { + "epoch": 124.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.65132265328175, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8636363636363636, \"recall\": 0.6129032258064516, \"f1-score\": 0.7169811320754716, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6829268292682927, \"recall\": 0.7567567567567568, \"f1-score\": 0.7179487179487181, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7806580614507443, \"recall\": 0.7996797200426232, \"f1-score\": 0.7847306037226593, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7553021200834489, \"recall\": 0.7413793103448276, \"f1-score\": 0.7403831446892091, \"support\": 116.0}}", + "eval_f1_macro": 48.29111407524058, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.0383144689209, + "eval_loss": 1.0762184858322144, + "eval_runtime": 2.2546, + "eval_samples_per_second": 51.451, + "step": 1984 + }, + { + "epoch": 125.0, + "learning_rate": 0.000125, + "loss": 0.0002, + "step": 2000 + }, + { + "epoch": 125.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0707623347613, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.7096774193548387, \"f1-score\": 0.7213114754098361, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.783531746031746, \"recall\": 0.7982629807226582, \"f1-score\": 0.7892417571882627, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7330459770114942, \"recall\": 0.7327586206896551, \"f1-score\": 0.7316260168183646, \"support\": 116.0}}", + "eval_f1_macro": 48.56872351927771, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.16260168183646, + "eval_loss": 1.0365558862686157, + "eval_runtime": 2.1837, + "eval_samples_per_second": 53.12, + "step": 2000 + }, + { + "epoch": 126.0, + "learning_rate": 0.0001233333333333333, + "loss": 0.0008, + "step": 2016 + }, + { + "epoch": 126.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.18828070182492, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7553631553631555, \"recall\": 0.8040873534825148, \"f1-score\": 0.7754215196479262, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7279218705942844, \"recall\": 0.7241379310344828, \"f1-score\": 0.7220728923745383, \"support\": 116.0}}", + "eval_f1_macro": 47.7182473629493, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.20728923745384, + "eval_loss": 1.0847417116165161, + "eval_runtime": 2.1565, + "eval_samples_per_second": 53.791, + "step": 2016 + }, + { + "epoch": 127.0, + "learning_rate": 0.00012166666666666665, + "loss": 0.0015, + "step": 2032 + }, + { + "epoch": 127.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.8450022809264, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.7419354838709677, \"f1-score\": 0.71875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7963338744588745, \"recall\": 0.8022952387871742, \"f1-score\": 0.7970214793913517, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.744126921928646, \"recall\": 0.7413793103448276, \"f1-score\": 0.7405667139988761, \"support\": 116.0}}", + "eval_f1_macro": 49.04747565485241, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.0566713998876, + "eval_loss": 1.0856082439422607, + "eval_runtime": 2.2011, + "eval_samples_per_second": 52.701, + "step": 2032 + }, + { + "epoch": 128.0, + "learning_rate": 0.00011999999999999999, + "loss": 0.0002, + "step": 2048 + }, + { + "epoch": 128.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.14339660631728, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.7096774193548387, \"f1-score\": 0.6984126984126984, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7924763655462185, \"recall\": 0.7982629807226582, \"f1-score\": 0.7932720732321743, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7347734352361635, \"recall\": 0.7327586206896551, \"f1-score\": 0.7320511931920426, \"support\": 116.0}}", + "eval_f1_macro": 48.81674296813381, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.20511931920429, + "eval_loss": 1.0866280794143677, + "eval_runtime": 2.1867, + "eval_samples_per_second": 53.047, + "step": 2048 + }, + { + "epoch": 129.0, + "learning_rate": 0.00011833333333333331, + "loss": 0.0002, + "step": 2064 + }, + { + "epoch": 129.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.85334504297987, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6756756756756757, \"f1-score\": 0.6944444444444444, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7962989631336406, \"recall\": 0.8016413591010365, \"f1-score\": 0.7969786766519763, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7433805418719213, \"recall\": 0.7413793103448276, \"f1-score\": 0.7409267646283232, \"support\": 116.0}}", + "eval_f1_macro": 49.044841640121625, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.09267646283233, + "eval_loss": 1.0848969221115112, + "eval_runtime": 2.0553, + "eval_samples_per_second": 56.44, + "step": 2064 + }, + { + "epoch": 130.0, + "learning_rate": 0.00011666666666666665, + "loss": 0.0002, + "step": 2080 + }, + { + "epoch": 130.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.18824972788626, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7967403307476837, \"recall\": 0.7982629807226582, \"f1-score\": 0.7957541397767409, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7348898777017439, \"recall\": 0.7327586206896551, \"f1-score\": 0.7323178924889144, \"support\": 116.0}}", + "eval_f1_macro": 48.969485524722515, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.23178924889146, + "eval_loss": 1.095829963684082, + "eval_runtime": 2.0853, + "eval_samples_per_second": 55.627, + "step": 2080 + }, + { + "epoch": 131.0, + "learning_rate": 0.000115, + "loss": 0.0005, + "step": 2096 + }, + { + "epoch": 131.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.29230917287957, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7733818645012864, \"recall\": 0.794230722658142, \"f1-score\": 0.7794353982112888, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7298702490941247, \"recall\": 0.7241379310344828, \"f1-score\": 0.7237639521008089, \"support\": 116.0}}", + "eval_f1_macro": 47.96525527454085, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.37639521008089, + "eval_loss": 1.105514407157898, + "eval_runtime": 2.1232, + "eval_samples_per_second": 54.634, + "step": 2096 + }, + { + "epoch": 132.0, + "learning_rate": 0.00011333333333333331, + "loss": 0.0002, + "step": 2112 + }, + { + "epoch": 132.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.80477325033718, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7911440451311141, \"recall\": 0.8009874794148988, \"f1-score\": 0.7933992764095822, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.744839157242546, \"recall\": 0.7413793103448276, \"f1-score\": 0.7411866007640886, \"support\": 116.0}}", + "eval_f1_macro": 48.82457085597429, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.11866007640886, + "eval_loss": 1.083550214767456, + "eval_runtime": 2.2453, + "eval_samples_per_second": 51.664, + "step": 2112 + }, + { + "epoch": 133.0, + "learning_rate": 0.00011166666666666667, + "loss": 0.0007, + "step": 2128 + }, + { + "epoch": 133.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.12366410068819, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.7027027027027027, \"f1-score\": 0.7123287671232876, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.756097582972583, \"recall\": 0.7929229632858665, \"f1-score\": 0.7695261745372544, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7315296437279195, \"recall\": 0.7241379310344828, \"f1-score\": 0.7231161330125597, \"support\": 116.0}}", + "eval_f1_macro": 47.355456894600266, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.31161330125597, + "eval_loss": 1.164588451385498, + "eval_runtime": 2.2443, + "eval_samples_per_second": 51.686, + "step": 2128 + }, + { + "epoch": 134.0, + "learning_rate": 0.00010999999999999998, + "loss": 0.0012, + "step": 2144 + }, + { + "epoch": 134.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.51241324001258, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.7096774193548387, \"f1-score\": 0.6769230769230768, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.6216216216216216, \"f1-score\": 0.676470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7628919651927243, \"recall\": 0.7844679356776131, \"f1-score\": 0.7716595882617554, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7192905962061232, \"recall\": 0.7155172413793104, \"f1-score\": 0.7145946079115717, \"support\": 116.0}}", + "eval_f1_macro": 47.486743893031104, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.45946079115717, + "eval_loss": 1.159348487854004, + "eval_runtime": 2.0817, + "eval_samples_per_second": 55.725, + "step": 2144 + }, + { + "epoch": 135.0, + "learning_rate": 0.00010833333333333333, + "loss": 0.0001, + "step": 2160 + }, + { + "epoch": 135.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.09424591991048, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.7419354838709677, \"f1-score\": 0.6969696969696969, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.6486486486486487, \"f1-score\": 0.6857142857142857, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7940569846819847, \"recall\": 0.7884063498982854, \"f1-score\": 0.7893731655341563, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.73658130662441, \"recall\": 0.7327586206896551, \"f1-score\": 0.732484493549936, \"support\": 116.0}}", + "eval_f1_macro": 48.5768101867173, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.24844935499361, + "eval_loss": 1.1159605979919434, + "eval_runtime": 2.0586, + "eval_samples_per_second": 56.35, + "step": 2160 + }, + { + "epoch": 136.0, + "learning_rate": 0.00010666666666666667, + "loss": 0.0008, + "step": 2176 + }, + { + "epoch": 136.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.66205711093811, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.717948717948718, \"recall\": 0.7567567567567568, \"f1-score\": 0.736842105263158, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7792376373626374, \"recall\": 0.7996797200426232, \"f1-score\": 0.7866254295240364, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7448609005936592, \"recall\": 0.7413793103448276, \"f1-score\": 0.739646476348462, \"support\": 116.0}}", + "eval_f1_macro": 48.4077187399407, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 73.9646476348462, + "eval_loss": 1.1159119606018066, + "eval_runtime": 2.2156, + "eval_samples_per_second": 52.355, + "step": 2176 + }, + { + "epoch": 137.0, + "learning_rate": 0.00010499999999999999, + "loss": 0.0002, + "step": 2192 + }, + { + "epoch": 137.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.57872110723775, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6486486486486487, \"f1-score\": 0.676056338028169, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7343265639589169, \"recall\": 0.7861662065291097, \"f1-score\": 0.7545112618900938, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7140724910653917, \"recall\": 0.7068965517241379, \"f1-score\": 0.7050411181396382, \"support\": 116.0}}", + "eval_f1_macro": 46.43146227015962, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.50411181396383, + "eval_loss": 1.1595603227615356, + "eval_runtime": 2.1661, + "eval_samples_per_second": 53.552, + "step": 2192 + }, + { + "epoch": 138.0, + "learning_rate": 0.00010333333333333333, + "loss": 0.0016, + "step": 2208 + }, + { + "epoch": 138.0, + "eval_accuracy": 68.96551724137932, + "eval_average_metrics": 63.281002597401425, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5405405405405406, \"f1-score\": 0.5970149253731343, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6896551724137931, \"macro avg\": {\"precision\": 0.748144216894217, \"recall\": 0.774332800542478, \"f1-score\": 0.75485509909281, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.6965268926475822, \"recall\": 0.6896551724137931, \"f1-score\": 0.6874035442421258, \"support\": 116.0}}", + "eval_f1_macro": 46.45262148263446, + "eval_f1_micro": 68.96551724137932, + "eval_f1_weighted": 68.74035442421258, + "eval_loss": 1.2053927183151245, + "eval_runtime": 2.062, + "eval_samples_per_second": 56.256, + "step": 2208 + }, + { + "epoch": 139.0, + "learning_rate": 0.00010166666666666667, + "loss": 0.0001, + "step": 2224 + }, + { + "epoch": 139.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 64.14137179915826, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7096774193548387, \"f1-score\": 0.6875, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6774193548387096, \"recall\": 0.5675675675675675, \"f1-score\": 0.6176470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7681467278241472, \"recall\": 0.7777111789208564, \"f1-score\": 0.7651180984580921, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7103061936209879, \"recall\": 0.6982758620689655, \"f1-score\": 0.6982612410849579, \"support\": 116.0}}", + "eval_f1_macro": 47.08419067434413, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.82612410849579, + "eval_loss": 1.1781898736953735, + "eval_runtime": 2.1597, + "eval_samples_per_second": 53.712, + "step": 2224 + }, + { + "epoch": 140.0, + "learning_rate": 9.999999999999999e-05, + "loss": 0.0011, + "step": 2240 + }, + { + "epoch": 140.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.48352678423673, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7619677197802198, \"recall\": 0.7965828731957764, \"f1-score\": 0.7700948561242679, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7300318933939624, \"recall\": 0.7155172413793104, \"f1-score\": 0.714402061765145, \"support\": 116.0}}", + "eval_f1_macro": 47.39045268457034, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.4402061765145, + "eval_loss": 1.1871286630630493, + "eval_runtime": 2.3012, + "eval_samples_per_second": 50.409, + "step": 2240 + }, + { + "epoch": 141.0, + "learning_rate": 9.833333333333333e-05, + "loss": 0.0003, + "step": 2256 + }, + { + "epoch": 141.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.6120488899911, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7768092814048697, \"recall\": 0.7908523442797636, \"f1-score\": 0.7754492703501312, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7285715678844482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7162479218563275, \"support\": 116.0}}", + "eval_f1_macro": 47.71995509846961, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.62479218563274, + "eval_loss": 1.1495442390441895, + "eval_runtime": 2.1903, + "eval_samples_per_second": 52.961, + "step": 2256 + }, + { + "epoch": 142.0, + "learning_rate": 9.666666666666667e-05, + "loss": 0.0009, + "step": 2272 + }, + { + "epoch": 142.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.6120488899911, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7768092814048697, \"recall\": 0.7908523442797636, \"f1-score\": 0.7754492703501312, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7285715678844482, \"recall\": 0.7155172413793104, \"f1-score\": 0.7162479218563275, \"support\": 116.0}}", + "eval_f1_macro": 47.71995509846961, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.62479218563274, + "eval_loss": 1.1490837335586548, + "eval_runtime": 2.1815, + "eval_samples_per_second": 53.174, + "step": 2272 + }, + { + "epoch": 143.0, + "learning_rate": 9.499999999999999e-05, + "loss": 0.0001, + "step": 2288 + }, + { + "epoch": 143.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.33454933252901, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7814369658119659, \"recall\": 0.794230722658142, \"f1-score\": 0.7793308787223594, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7389433602795671, \"recall\": 0.7241379310344828, \"f1-score\": 0.7255178781722812, \"support\": 116.0}}", + "eval_f1_macro": 47.95882330599135, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.55178781722812, + "eval_loss": 1.153124451637268, + "eval_runtime": 2.1286, + "eval_samples_per_second": 54.495, + "step": 2288 + }, + { + "epoch": 144.0, + "learning_rate": 9.333333333333333e-05, + "loss": 0.004, + "step": 2304 + }, + { + "epoch": 144.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 64.13153922628965, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6666666666666666, \"recall\": 0.8333333333333334, \"f1-score\": 0.7407407407407408, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.65625, \"recall\": 0.5675675675675675, \"f1-score\": 0.6086956521739131, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7650965683023873, \"recall\": 0.7840955875230069, \"f1-score\": 0.7655060069702013, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7102843741425043, \"recall\": 0.6982758620689655, \"f1-score\": 0.6976292252396847, \"support\": 116.0}}", + "eval_f1_macro": 47.108061967397, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.76292252396847, + "eval_loss": 1.1598789691925049, + "eval_runtime": 2.2008, + "eval_samples_per_second": 52.709, + "step": 2304 + }, + { + "epoch": 145.0, + "learning_rate": 9.166666666666667e-05, + "loss": 0.0009, + "step": 2320 + }, + { + "epoch": 145.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.55450050368793, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6470588235294118, \"recall\": 0.9166666666666666, \"f1-score\": 0.7586206896551724, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.8, \"recall\": 0.6451612903225806, \"f1-score\": 0.7142857142857142, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.696969696969697, \"recall\": 0.6216216216216216, \"f1-score\": 0.6571428571428571, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.8888888888888888, \"f1-score\": 0.6956521739130435, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7717238031576267, \"recall\": 0.7972367528819141, \"f1-score\": 0.771630046321377, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7381072570093868, \"recall\": 0.7155172413793104, \"f1-score\": 0.7162962781142028, \"support\": 116.0}}", + "eval_f1_macro": 47.48492592746936, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.62962781142028, + "eval_loss": 1.2059998512268066, + "eval_runtime": 2.077, + "eval_samples_per_second": 55.849, + "step": 2320 + }, + { + "epoch": 146.0, + "learning_rate": 8.999999999999999e-05, + "loss": 0.0008, + "step": 2336 + }, + { + "epoch": 146.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.68990280670708, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7916666666666666, \"recall\": 0.6129032258064516, \"f1-score\": 0.6909090909090909, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7486594798175681, \"recall\": 0.793204494817398, \"f1-score\": 0.7620284840355263, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7195403691980772, \"recall\": 0.7068965517241379, \"f1-score\": 0.7048624032596834, \"support\": 116.0}}", + "eval_f1_macro": 46.894060556032386, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.48624032596834, + "eval_loss": 1.19475519657135, + "eval_runtime": 2.2669, + "eval_samples_per_second": 51.171, + "step": 2336 + }, + { + "epoch": 147.0, + "learning_rate": 8.833333333333333e-05, + "loss": 0.0002, + "step": 2352 + }, + { + "epoch": 147.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.70775012641597, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.7096774193548387, \"f1-score\": 0.7096774193548389, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.5675675675675675, \"f1-score\": 0.626865671641791, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7515683465280241, \"recall\": 0.788127845587523, \"f1-score\": 0.7635325328392253, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7135894278135657, \"recall\": 0.7068965517241379, \"f1-score\": 0.7046507275534553, \"support\": 116.0}}", + "eval_f1_macro": 46.98661740549079, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.46507275534553, + "eval_loss": 1.1726635694503784, + "eval_runtime": 2.2233, + "eval_samples_per_second": 52.176, + "step": 2352 + }, + { + "epoch": 148.0, + "learning_rate": 8.666666666666665e-05, + "loss": 0.0001, + "step": 2368 + }, + { + "epoch": 148.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 64.00347575303101, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6774193548387096, \"recall\": 0.5675675675675675, \"f1-score\": 0.6176470588235294, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7475360884635078, \"recall\": 0.7840955875230069, \"f1-score\": 0.7597361025880794, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7038007737534989, \"recall\": 0.6982758620689655, \"f1-score\": 0.6960573966983371, \"support\": 116.0}}", + "eval_f1_macro": 46.7529909284972, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.60573966983371, + "eval_loss": 1.1674237251281738, + "eval_runtime": 2.108, + "eval_samples_per_second": 55.028, + "step": 2368 + }, + { + "epoch": 149.0, + "learning_rate": 8.499999999999999e-05, + "loss": 0.0013, + "step": 2384 + }, + { + "epoch": 149.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.57406977941876, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7715246354952238, \"recall\": 0.7908523442797636, \"f1-score\": 0.7735764699479191, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.726712435637385, \"recall\": 0.7155172413793104, \"f1-score\": 0.7158812499886412, \"support\": 116.0}}", + "eval_f1_macro": 47.60470584294886, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.58812499886412, + "eval_loss": 1.1712332963943481, + "eval_runtime": 2.1727, + "eval_samples_per_second": 53.391, + "step": 2384 + }, + { + "epoch": 150.0, + "learning_rate": 8.333333333333333e-05, + "loss": 0.0006, + "step": 2400 + }, + { + "epoch": 150.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.15508351872047, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.8181818181818182, \"recall\": 0.75, \"f1-score\": 0.7826086956521738, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.7741935483870968, \"f1-score\": 0.7164179104477612, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.71875, \"recall\": 0.6216216216216216, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.8022174873737374, \"recall\": 0.7925324518066453, \"f1-score\": 0.7928398963556414, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7414299242424242, \"recall\": 0.7327586206896551, \"f1-score\": 0.7327846246891143, \"support\": 116.0}}", + "eval_f1_macro": 48.79014746803947, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.27846246891143, + "eval_loss": 1.21891450881958, + "eval_runtime": 2.3304, + "eval_samples_per_second": 49.776, + "step": 2400 + }, + { + "epoch": 151.0, + "learning_rate": 8.166666666666665e-05, + "loss": 0.0059, + "step": 2416 + }, + { + "epoch": 151.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.85326013055257, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7978805728805729, \"recall\": 0.8009874794148988, \"f1-score\": 0.794782459278445, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7494031830238728, \"recall\": 0.7413793103448276, \"f1-score\": 0.7422748865149431, \"support\": 116.0}}", + "eval_f1_macro": 48.90968980175047, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.22748865149431, + "eval_loss": 1.139691948890686, + "eval_runtime": 2.1983, + "eval_samples_per_second": 52.768, + "step": 2416 + }, + { + "epoch": 152.0, + "learning_rate": 7.999999999999999e-05, + "loss": 0.0004, + "step": 2432 + }, + { + "epoch": 152.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.85326013055257, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.7978805728805729, \"recall\": 0.8009874794148988, \"f1-score\": 0.794782459278445, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7494031830238728, \"recall\": 0.7413793103448276, \"f1-score\": 0.7422748865149431, \"support\": 116.0}}", + "eval_f1_macro": 48.90968980175047, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.22748865149431, + "eval_loss": 1.1354390382766724, + "eval_runtime": 2.0409, + "eval_samples_per_second": 56.838, + "step": 2432 + }, + { + "epoch": 153.0, + "learning_rate": 7.833333333333333e-05, + "loss": 0.0003, + "step": 2448 + }, + { + "epoch": 153.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.10759310976837, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7899801587301587, \"recall\": 0.7976091010365204, \"f1-score\": 0.7887149059506178, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7410851122058019, \"recall\": 0.7327586206896551, \"f1-score\": 0.7334234639648906, \"support\": 116.0}}", + "eval_f1_macro": 48.5363019046534, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.34234639648905, + "eval_loss": 1.137083888053894, + "eval_runtime": 1.9707, + "eval_samples_per_second": 58.861, + "step": 2448 + }, + { + "epoch": 154.0, + "learning_rate": 7.666666666666666e-05, + "loss": 0.0006, + "step": 2464 + }, + { + "epoch": 154.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.1067536864051, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.6875, \"recall\": 0.9166666666666666, \"f1-score\": 0.7857142857142857, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7890128968253968, \"recall\": 0.8046473893248086, \"f1-score\": 0.7900264310200789, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7429529282977559, \"recall\": 0.7327586206896551, \"f1-score\": 0.732582794679922, \"support\": 116.0}}", + "eval_f1_macro": 48.61701113969717, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.25827946799221, + "eval_loss": 1.142392873764038, + "eval_runtime": 2.2302, + "eval_samples_per_second": 52.013, + "step": 2464 + }, + { + "epoch": 155.0, + "learning_rate": 7.5e-05, + "loss": 0.0002, + "step": 2480 + }, + { + "epoch": 155.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.10759310976837, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7899801587301587, \"recall\": 0.7976091010365204, \"f1-score\": 0.7887149059506178, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7410851122058019, \"recall\": 0.7327586206896551, \"f1-score\": 0.7334234639648906, \"support\": 116.0}}", + "eval_f1_macro": 48.5363019046534, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.34234639648905, + "eval_loss": 1.127716064453125, + "eval_runtime": 1.9562, + "eval_samples_per_second": 59.298, + "step": 2480 + }, + { + "epoch": 156.0, + "learning_rate": 7.333333333333332e-05, + "loss": 0.0001, + "step": 2496 + }, + { + "epoch": 156.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.14425762652212, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7936155319776009, \"recall\": 0.7976091010365204, \"f1-score\": 0.7910780089493139, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7398576684248503, \"recall\": 0.7327586206896551, \"f1-score\": 0.7334358274050738, \"support\": 116.0}}", + "eval_f1_macro": 48.681723627650086, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.34358274050739, + "eval_loss": 1.127997875213623, + "eval_runtime": 2.0527, + "eval_samples_per_second": 56.51, + "step": 2496 + }, + { + "epoch": 157.0, + "learning_rate": 7.166666666666667e-05, + "loss": 0.0002, + "step": 2512 + }, + { + "epoch": 157.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.4337163915427, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7895070207570207, \"recall\": 0.794230722658142, \"f1-score\": 0.7873604776969747, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7306223948465328, \"recall\": 0.7241379310344828, \"f1-score\": 0.7245432688561426, \"support\": 116.0}}", + "eval_f1_macro": 48.452952473659984, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.45432688561426, + "eval_loss": 1.1359306573867798, + "eval_runtime": 2.1072, + "eval_samples_per_second": 55.049, + "step": 2512 + }, + { + "epoch": 158.0, + "learning_rate": 7e-05, + "loss": 0.0001, + "step": 2528 + }, + { + "epoch": 158.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 64.12329403555103, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.65625, \"recall\": 0.5675675675675675, \"f1-score\": 0.6086956521739131, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.762747061965812, \"recall\": 0.7840955875230069, \"f1-score\": 0.7657600338997346, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7069006962864721, \"recall\": 0.6982758620689655, \"f1-score\": 0.6971430933458117, \"support\": 116.0}}", + "eval_f1_macro": 47.123694393829815, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.71430933458117, + "eval_loss": 1.1599950790405273, + "eval_runtime": 2.2991, + "eval_samples_per_second": 50.454, + "step": 2528 + }, + { + "epoch": 159.0, + "learning_rate": 6.833333333333333e-05, + "loss": 0.0003, + "step": 2544 + }, + { + "epoch": 159.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.84977905236525, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7670663866784557, \"recall\": 0.7874739659013852, \"f1-score\": 0.7696789321789321, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7166739077410896, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065494849977608, \"support\": 116.0}}", + "eval_f1_macro": 47.36485736485736, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.65494849977608, + "eval_loss": 1.1744282245635986, + "eval_runtime": 2.0565, + "eval_samples_per_second": 56.406, + "step": 2544 + }, + { + "epoch": 160.0, + "learning_rate": 6.666666666666666e-05, + "loss": 0.0004, + "step": 2560 + }, + { + "epoch": 160.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.8131145356115, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7634310134310134, \"recall\": 0.7874739659013852, \"f1-score\": 0.767315829180236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7179013515220412, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065371215575774, \"support\": 116.0}}", + "eval_f1_macro": 47.21943564186068, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.65371215575774, + "eval_loss": 1.1818971633911133, + "eval_runtime": 2.324, + "eval_samples_per_second": 49.915, + "step": 2560 + }, + { + "epoch": 161.0, + "learning_rate": 6.5e-05, + "loss": 0.0001, + "step": 2576 + }, + { + "epoch": 161.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.54092535499936, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7681287258493141, \"recall\": 0.7908523442797636, \"f1-score\": 0.7712645066336179, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7284518360836819, \"recall\": 0.7155172413793104, \"f1-score\": 0.7159782196668197, \"support\": 116.0}}", + "eval_f1_macro": 47.462431177453404, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.59782196668198, + "eval_loss": 1.2029889822006226, + "eval_runtime": 2.1328, + "eval_samples_per_second": 54.388, + "step": 2576 + }, + { + "epoch": 162.0, + "learning_rate": 6.333333333333333e-05, + "loss": 0.0003, + "step": 2592 + }, + { + "epoch": 162.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.70888807942681, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7492032069237953, \"recall\": 0.7868200862152475, \"f1-score\": 0.7613237063570644, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7163155049818335, \"recall\": 0.7068965517241379, \"f1-score\": 0.706055523509065, \"support\": 116.0}}", + "eval_f1_macro": 46.85068962197319, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.60555235090649, + "eval_loss": 1.2051033973693848, + "eval_runtime": 2.3524, + "eval_samples_per_second": 49.312, + "step": 2592 + }, + { + "epoch": 163.0, + "learning_rate": 6.166666666666666e-05, + "loss": 0.0002, + "step": 2608 + }, + { + "epoch": 163.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.70888807942681, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.6451612903225806, \"f1-score\": 0.7017543859649122, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6764705882352942, \"recall\": 0.6216216216216216, \"f1-score\": 0.6478873239436619, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7492032069237953, \"recall\": 0.7868200862152475, \"f1-score\": 0.7613237063570644, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7163155049818335, \"recall\": 0.7068965517241379, \"f1-score\": 0.706055523509065, \"support\": 116.0}}", + "eval_f1_macro": 46.85068962197319, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.60555235090649, + "eval_loss": 1.1949589252471924, + "eval_runtime": 2.0434, + "eval_samples_per_second": 56.768, + "step": 2608 + }, + { + "epoch": 164.0, + "learning_rate": 5.9999999999999995e-05, + "loss": 0.0, + "step": 2624 + }, + { + "epoch": 164.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.8131145356115, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7634310134310134, \"recall\": 0.7874739659013852, \"f1-score\": 0.767315829180236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7179013515220412, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065371215575774, \"support\": 116.0}}", + "eval_f1_macro": 47.21943564186068, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.65371215575774, + "eval_loss": 1.189142107963562, + "eval_runtime": 1.9698, + "eval_samples_per_second": 58.89, + "step": 2624 + }, + { + "epoch": 165.0, + "learning_rate": 5.8333333333333326e-05, + "loss": 0.0004, + "step": 2640 + }, + { + "epoch": 165.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.982577048752844, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", + "eval_f1_macro": 46.6090349379474, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.66610084327087, + "eval_loss": 1.203819751739502, + "eval_runtime": 2.1425, + "eval_samples_per_second": 54.142, + "step": 2640 + }, + { + "epoch": 166.0, + "learning_rate": 5.666666666666666e-05, + "loss": 0.0001, + "step": 2656 + }, + { + "epoch": 166.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.982577048752844, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", + "eval_f1_macro": 46.6090349379474, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.66610084327087, + "eval_loss": 1.2178500890731812, + "eval_runtime": 2.0767, + "eval_samples_per_second": 55.858, + "step": 2656 + }, + { + "epoch": 167.0, + "learning_rate": 5.499999999999999e-05, + "loss": 0.0001, + "step": 2672 + }, + { + "epoch": 167.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.982577048752844, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", + "eval_f1_macro": 46.6090349379474, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.66610084327087, + "eval_loss": 1.2182705402374268, + "eval_runtime": 1.9593, + "eval_samples_per_second": 59.206, + "step": 2672 + }, + { + "epoch": 168.0, + "learning_rate": 5.333333333333333e-05, + "loss": 0.0001, + "step": 2688 + }, + { + "epoch": 168.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.982577048752844, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", + "eval_f1_macro": 46.6090349379474, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.66610084327087, + "eval_loss": 1.2183611392974854, + "eval_runtime": 2.3149, + "eval_samples_per_second": 50.11, + "step": 2688 + }, + { + "epoch": 169.0, + "learning_rate": 5.1666666666666664e-05, + "loss": 0.0001, + "step": 2704 + }, + { + "epoch": 169.0, + "eval_accuracy": 69.82758620689656, + "eval_average_metrics": 63.982577048752844, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.6451612903225806, \"f1-score\": 0.689655172413793, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.6982758620689655, \"macro avg\": {\"precision\": 0.7444164631664631, \"recall\": 0.7834417078368692, \"f1-score\": 0.7573968177416452, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7055746775574362, \"recall\": 0.6982758620689655, \"f1-score\": 0.6966610084327087, \"support\": 116.0}}", + "eval_f1_macro": 46.6090349379474, + "eval_f1_micro": 69.82758620689656, + "eval_f1_weighted": 69.66610084327087, + "eval_loss": 1.216108798980713, + "eval_runtime": 2.042, + "eval_samples_per_second": 56.806, + "step": 2704 + }, + { + "epoch": 170.0, + "learning_rate": 4.9999999999999996e-05, + "loss": 0.0001, + "step": 2720 + }, + { + "epoch": 170.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.8131145356115, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7634310134310134, \"recall\": 0.7874739659013852, \"f1-score\": 0.767315829180236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7179013515220412, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065371215575774, \"support\": 116.0}}", + "eval_f1_macro": 47.21943564186068, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.65371215575774, + "eval_loss": 1.203769326210022, + "eval_runtime": 2.0637, + "eval_samples_per_second": 56.21, + "step": 2720 + }, + { + "epoch": 171.0, + "learning_rate": 4.8333333333333334e-05, + "loss": 0.0001, + "step": 2736 + }, + { + "epoch": 171.0, + "eval_accuracy": 70.6896551724138, + "eval_average_metrics": 64.8131145356115, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6666666666666666, \"recall\": 0.5945945945945946, \"f1-score\": 0.6285714285714286, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"accuracy\": 0.7068965517241379, \"macro avg\": {\"precision\": 0.7634310134310134, \"recall\": 0.7874739659013852, \"f1-score\": 0.767315829180236, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7179013515220412, \"recall\": 0.7068965517241379, \"f1-score\": 0.7065371215575774, \"support\": 116.0}}", + "eval_f1_macro": 47.21943564186068, + "eval_f1_micro": 70.6896551724138, + "eval_f1_weighted": 70.65371215575774, + "eval_loss": 1.1848161220550537, + "eval_runtime": 2.1122, + "eval_samples_per_second": 54.919, + "step": 2736 + }, + { + "epoch": 172.0, + "learning_rate": 4.6666666666666665e-05, + "loss": 0.0001, + "step": 2752 + }, + { + "epoch": 172.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.89123924112491, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.8031652187902187, \"recall\": 0.8009874794148988, \"f1-score\": 0.7966552596806572, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.751262315270936, \"recall\": 0.7413793103448276, \"f1-score\": 0.7426415583826294, \"support\": 116.0}}", + "eval_f1_macro": 49.02493905727121, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.26415583826291, + "eval_loss": 1.1390397548675537, + "eval_runtime": 2.0345, + "eval_samples_per_second": 57.017, + "step": 2752 + }, + { + "epoch": 173.0, + "learning_rate": 4.4999999999999996e-05, + "loss": 0.0001, + "step": 2768 + }, + { + "epoch": 173.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.89123924112491, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.8031652187902187, \"recall\": 0.8009874794148988, \"f1-score\": 0.7966552596806572, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.751262315270936, \"recall\": 0.7413793103448276, \"f1-score\": 0.7426415583826294, \"support\": 116.0}}", + "eval_f1_macro": 49.02493905727121, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.26415583826291, + "eval_loss": 1.1333969831466675, + "eval_runtime": 2.0924, + "eval_samples_per_second": 55.44, + "step": 2768 + }, + { + "epoch": 174.0, + "learning_rate": 4.333333333333333e-05, + "loss": 0.0004, + "step": 2784 + }, + { + "epoch": 174.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.92790375787868, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.806800592037661, \"recall\": 0.8009874794148988, \"f1-score\": 0.7990183626793534, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7500348714899845, \"recall\": 0.7413793103448276, \"f1-score\": 0.7426539218228126, \"support\": 116.0}}", + "eval_f1_macro": 49.1703607802679, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.26539218228126, + "eval_loss": 1.1344233751296997, + "eval_runtime": 2.0203, + "eval_samples_per_second": 57.417, + "step": 2784 + }, + { + "epoch": 175.0, + "learning_rate": 4.1666666666666665e-05, + "loss": 0.0002, + "step": 2800 + }, + { + "epoch": 175.0, + "eval_accuracy": 74.13793103448276, + "eval_average_metrics": 67.92790375787868, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7692307692307693, \"recall\": 0.8333333333333334, \"f1-score\": 0.8, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.6774193548387096, \"f1-score\": 0.7, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7413793103448276, \"macro avg\": {\"precision\": 0.806800592037661, \"recall\": 0.8009874794148988, \"f1-score\": 0.7990183626793534, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7500348714899845, \"recall\": 0.7413793103448276, \"f1-score\": 0.7426539218228126, \"support\": 116.0}}", + "eval_f1_macro": 49.1703607802679, + "eval_f1_micro": 74.13793103448276, + "eval_f1_weighted": 74.26539218228126, + "eval_loss": 1.136767864227295, + "eval_runtime": 1.964, + "eval_samples_per_second": 59.062, + "step": 2800 + }, + { + "epoch": 176.0, + "learning_rate": 3.9999999999999996e-05, + "loss": 0.0001, + "step": 2816 + }, + { + "epoch": 176.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.38902324812496, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881944444444444, \"recall\": 0.7871924343698538, \"f1-score\": 0.7833935827198057, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7314176245210728, \"recall\": 0.7241379310344828, \"f1-score\": 0.7251967092592287, \"support\": 116.0}}", + "eval_f1_macro": 48.208835859680356, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.51967092592288, + "eval_loss": 1.1415574550628662, + "eval_runtime": 1.9714, + "eval_samples_per_second": 58.841, + "step": 2816 + }, + { + "epoch": 177.0, + "learning_rate": 3.833333333333333e-05, + "loss": 0.0, + "step": 2832 + }, + { + "epoch": 177.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.5615084791178, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.6451612903225806, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7690441981390257, \"recall\": 0.7831601763053376, \"f1-score\": 0.7735184851397902, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7188008417794387, \"recall\": 0.7155172413793104, \"f1-score\": 0.7154144809354516, \"support\": 116.0}}", + "eval_f1_macro": 47.601137547064006, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.54144809354516, + "eval_loss": 1.1445733308792114, + "eval_runtime": 1.9924, + "eval_samples_per_second": 58.22, + "step": 2832 + }, + { + "epoch": 178.0, + "learning_rate": 3.666666666666666e-05, + "loss": 0.0, + "step": 2848 + }, + { + "epoch": 178.0, + "eval_accuracy": 71.55172413793103, + "eval_average_metrics": 65.5615084791178, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.8571428571428571, \"recall\": 0.75, \"f1-score\": 0.7999999999999999, \"support\": 8.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.6451612903225806, \"f1-score\": 0.6666666666666667, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7155172413793104, \"macro avg\": {\"precision\": 0.7690441981390257, \"recall\": 0.7831601763053376, \"f1-score\": 0.7735184851397902, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7188008417794387, \"recall\": 0.7155172413793104, \"f1-score\": 0.7154144809354516, \"support\": 116.0}}", + "eval_f1_macro": 47.601137547064006, + "eval_f1_micro": 71.55172413793103, + "eval_f1_weighted": 71.54144809354516, + "eval_loss": 1.144472360610962, + "eval_runtime": 2.2321, + "eval_samples_per_second": 51.969, + "step": 2848 + }, + { + "epoch": 179.0, + "learning_rate": 3.5e-05, + "loss": 0.0098, + "step": 2864 + }, + { + "epoch": 179.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.38902324812496, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881944444444444, \"recall\": 0.7871924343698538, \"f1-score\": 0.7833935827198057, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7314176245210728, \"recall\": 0.7241379310344828, \"f1-score\": 0.7251967092592287, \"support\": 116.0}}", + "eval_f1_macro": 48.208835859680356, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.51967092592288, + "eval_loss": 1.142303705215454, + "eval_runtime": 1.9646, + "eval_samples_per_second": 59.044, + "step": 2864 + }, + { + "epoch": 180.0, + "learning_rate": 3.333333333333333e-05, + "loss": 0.0001, + "step": 2880 + }, + { + "epoch": 180.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.38902324812496, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.6774193548387096, \"f1-score\": 0.6885245901639343, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7881944444444444, \"recall\": 0.7871924343698538, \"f1-score\": 0.7833935827198057, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7314176245210728, \"recall\": 0.7241379310344828, \"f1-score\": 0.7251967092592287, \"support\": 116.0}}", + "eval_f1_macro": 48.208835859680356, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.51967092592288, + "eval_loss": 1.133960485458374, + "eval_runtime": 2.0714, + "eval_samples_per_second": 56.001, + "step": 2880 + }, + { + "epoch": 181.0, + "learning_rate": 3.1666666666666666e-05, + "loss": 0.0001, + "step": 2896 + }, + { + "epoch": 181.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.95417031249403, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.7297297297297297, \"f1-score\": 0.7200000000000001, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7741431704260652, \"recall\": 0.7963013416642448, \"f1-score\": 0.7812629424239331, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7391888125486129, \"recall\": 0.7327586206896551, \"f1-score\": 0.7318723757826459, \"support\": 116.0}}", + "eval_f1_macro": 48.0777195337805, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.18723757826459, + "eval_loss": 1.1398240327835083, + "eval_runtime": 2.1727, + "eval_samples_per_second": 53.39, + "step": 2896 + }, + { + "epoch": 182.0, + "learning_rate": 2.9999999999999997e-05, + "loss": 0.0006, + "step": 2912 + }, + { + "epoch": 182.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.95417031249403, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.7297297297297297, \"f1-score\": 0.7200000000000001, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7741431704260652, \"recall\": 0.7963013416642448, \"f1-score\": 0.7812629424239331, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7391888125486129, \"recall\": 0.7327586206896551, \"f1-score\": 0.7318723757826459, \"support\": 116.0}}", + "eval_f1_macro": 48.0777195337805, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.18723757826459, + "eval_loss": 1.1363673210144043, + "eval_runtime": 2.1014, + "eval_samples_per_second": 55.2, + "step": 2912 + }, + { + "epoch": 183.0, + "learning_rate": 2.833333333333333e-05, + "loss": 0.0016, + "step": 2928 + }, + { + "epoch": 183.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 66.95417031249403, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.7297297297297297, \"f1-score\": 0.7200000000000001, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7741431704260652, \"recall\": 0.7963013416642448, \"f1-score\": 0.7812629424239331, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7391888125486129, \"recall\": 0.7327586206896551, \"f1-score\": 0.7318723757826459, \"support\": 116.0}}", + "eval_f1_macro": 48.0777195337805, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.18723757826459, + "eval_loss": 1.1402283906936646, + "eval_runtime": 2.0837, + "eval_samples_per_second": 55.671, + "step": 2928 + }, + { + "epoch": 184.0, + "learning_rate": 2.6666666666666667e-05, + "loss": 0.0021, + "step": 2944 + }, + { + "epoch": 184.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.18929941059953, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.75, \"f1-score\": 0.75, \"support\": 8.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.6129032258064516, \"f1-score\": 0.6785714285714285, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.7027027027027027, \"f1-score\": 0.7027027027027027, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.7640743096993097, \"recall\": 0.7929229632858665, \"f1-score\": 0.7738376223670341, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7310507165248544, \"recall\": 0.7241379310344828, \"f1-score\": 0.7230883467445334, \"support\": 116.0}}", + "eval_f1_macro": 47.620776761048255, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.30883467445334, + "eval_loss": 1.145507574081421, + "eval_runtime": 2.1519, + "eval_samples_per_second": 53.906, + "step": 2944 + }, + { + "epoch": 185.0, + "learning_rate": 2.4999999999999998e-05, + "loss": 0.0, + "step": 2960 + }, + { + "epoch": 185.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0801291345566, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", + "eval_f1_macro": 48.36572234302737, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.40307005726802, + "eval_loss": 1.1441426277160645, + "eval_runtime": 2.286, + "eval_samples_per_second": 50.743, + "step": 2960 + }, + { + "epoch": 186.0, + "learning_rate": 2.3333333333333332e-05, + "loss": 0.0, + "step": 2976 + }, + { + "epoch": 186.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0801291345566, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", + "eval_f1_macro": 48.36572234302737, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.40307005726802, + "eval_loss": 1.1459873914718628, + "eval_runtime": 2.0524, + "eval_samples_per_second": 56.519, + "step": 2976 + }, + { + "epoch": 187.0, + "learning_rate": 2.1666666666666664e-05, + "loss": 0.0001, + "step": 2992 + }, + { + "epoch": 187.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0801291345566, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", + "eval_f1_macro": 48.36572234302737, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.40307005726802, + "eval_loss": 1.148589849472046, + "eval_runtime": 2.3314, + "eval_samples_per_second": 49.755, + "step": 2992 + }, + { + "epoch": 188.0, + "learning_rate": 1.9999999999999998e-05, + "loss": 0.0006, + "step": 3008 + }, + { + "epoch": 188.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0801291345566, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", + "eval_f1_macro": 48.36572234302737, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.40307005726802, + "eval_loss": 1.1534448862075806, + "eval_runtime": 2.0097, + "eval_samples_per_second": 57.72, + "step": 3008 + }, + { + "epoch": 189.0, + "learning_rate": 1.833333333333333e-05, + "loss": 0.0001, + "step": 3024 + }, + { + "epoch": 189.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.0801291345566, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.6774193548387096, \"f1-score\": 0.7241379310344828, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7889384920634921, \"recall\": 0.7976091010365204, \"f1-score\": 0.7859429880741948, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7457067597153804, \"recall\": 0.7327586206896551, \"f1-score\": 0.7340307005726803, \"support\": 116.0}}", + "eval_f1_macro": 48.36572234302737, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.40307005726802, + "eval_loss": 1.1526384353637695, + "eval_runtime": 2.1314, + "eval_samples_per_second": 54.425, + "step": 3024 + }, + { + "epoch": 190.0, + "learning_rate": 1.6666666666666664e-05, + "loss": 0.0001, + "step": 3040 + }, + { + "epoch": 190.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.1524851322174072, + "eval_runtime": 2.1944, + "eval_samples_per_second": 52.862, + "step": 3040 + }, + { + "epoch": 191.0, + "learning_rate": 1.4999999999999999e-05, + "loss": 0.0001, + "step": 3056 + }, + { + "epoch": 191.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.12167806340173, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7930420274170275, \"recall\": 0.7976091010365204, \"f1-score\": 0.7891707023042469, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7429855824252376, \"recall\": 0.7327586206896551, \"f1-score\": 0.7337063720464534, \"support\": 116.0}}", + "eval_f1_macro": 48.56435091103058, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.37063720464533, + "eval_loss": 1.142797589302063, + "eval_runtime": 1.9594, + "eval_samples_per_second": 59.201, + "step": 3056 + }, + { + "epoch": 192.0, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.0002, + "step": 3072 + }, + { + "epoch": 192.0, + "eval_accuracy": 73.27586206896551, + "eval_average_metrics": 67.12167806340173, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6756756756756757, \"f1-score\": 0.684931506849315, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"accuracy\": 0.7327586206896551, \"macro avg\": {\"precision\": 0.7930420274170275, \"recall\": 0.7976091010365204, \"f1-score\": 0.7891707023042469, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7429855824252376, \"recall\": 0.7327586206896551, \"f1-score\": 0.7337063720464534, \"support\": 116.0}}", + "eval_f1_macro": 48.56435091103058, + "eval_f1_micro": 73.27586206896551, + "eval_f1_weighted": 73.37063720464533, + "eval_loss": 1.14202880859375, + "eval_runtime": 2.1118, + "eval_samples_per_second": 54.931, + "step": 3072 + }, + { + "epoch": 193.0, + "learning_rate": 1.1666666666666666e-05, + "loss": 0.0, + "step": 3088 + }, + { + "epoch": 193.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.1449861526489258, + "eval_runtime": 2.0551, + "eval_samples_per_second": 56.444, + "step": 3088 + }, + { + "epoch": 194.0, + "learning_rate": 9.999999999999999e-06, + "loss": 0.0, + "step": 3104 + }, + { + "epoch": 194.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.1457918882369995, + "eval_runtime": 2.0705, + "eval_samples_per_second": 56.025, + "step": 3104 + }, + { + "epoch": 195.0, + "learning_rate": 8.333333333333332e-06, + "loss": 0.0006, + "step": 3120 + }, + { + "epoch": 195.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.1468638181686401, + "eval_runtime": 1.9389, + "eval_samples_per_second": 59.828, + "step": 3120 + }, + { + "epoch": 196.0, + "learning_rate": 6.666666666666667e-06, + "loss": 0.0005, + "step": 3136 + }, + { + "epoch": 196.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.150787591934204, + "eval_runtime": 2.0877, + "eval_samples_per_second": 55.564, + "step": 3136 + }, + { + "epoch": 197.0, + "learning_rate": 4.9999999999999996e-06, + "loss": 0.0001, + "step": 3152 + }, + { + "epoch": 197.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.1514496803283691, + "eval_runtime": 2.0764, + "eval_samples_per_second": 55.866, + "step": 3152 + }, + { + "epoch": 198.0, + "learning_rate": 3.3333333333333333e-06, + "loss": 0.0001, + "step": 3168 + }, + { + "epoch": 198.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.1519808769226074, + "eval_runtime": 2.0667, + "eval_samples_per_second": 56.128, + "step": 3168 + }, + { + "epoch": 199.0, + "learning_rate": 1.6666666666666667e-06, + "loss": 0.0001, + "step": 3184 + }, + { + "epoch": 199.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.1532700061798096, + "eval_runtime": 2.077, + "eval_samples_per_second": 55.849, + "step": 3184 + }, + { + "epoch": 200.0, + "learning_rate": 0.0, + "loss": 0.0001, + "step": 3200 + }, + { + "epoch": 200.0, + "eval_accuracy": 72.41379310344827, + "eval_average_metrics": 66.3627208342036, + "eval_classification_report": "{\"accusation\": {\"precision\": 0.7142857142857143, \"recall\": 0.8333333333333334, \"f1-score\": 0.7692307692307692, \"support\": 12.0}, \"appreciation\": {\"precision\": 1.0, \"recall\": 0.75, \"f1-score\": 0.8571428571428571, \"support\": 8.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6774193548387096, \"f1-score\": 0.7118644067796611, \"support\": 31.0}, \"evaluation\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6486486486486487, \"f1-score\": 0.6666666666666667, \"support\": 37.0}, \"question\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 1.0}, \"rejection\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.8888888888888888, \"f1-score\": 0.761904761904762, \"support\": 9.0}, \"accuracy\": 0.7241379310344828, \"macro avg\": {\"precision\": 0.784375, \"recall\": 0.794230722658142, \"f1-score\": 0.782125692519511, \"support\": 116.0}, \"weighted avg\": {\"precision\": 0.7354987684729063, \"recall\": 0.7241379310344828, \"f1-score\": 0.7249248528256333, \"support\": 116.0}}", + "eval_f1_macro": 48.13081184735453, + "eval_f1_micro": 72.41379310344828, + "eval_f1_weighted": 72.49248528256334, + "eval_loss": 1.1530934572219849, + "eval_runtime": 2.0419, + "eval_samples_per_second": 56.811, + "step": 3200 + }, + { + "epoch": 200.0, + "step": 3200, + "total_flos": 1.5716799379144704e+16, + "train_runtime": 2386.5323, + "train_samples_per_second": 1.341 + } + ], + "max_steps": 3200, + "num_train_epochs": 200, + "total_flos": 1.5716799379144704e+16, + "trial_name": null, + "trial_params": null +}