diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,3823 @@ +{ + "best_metric": 77.22228406219811, + "best_model_checkpoint": "/mounts/work/faeze/attempt/new_setting_outputs/source_adapter/crisis_8/100/checkpoint-1479", + "epoch": 200.0, + "global_step": 3400, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 1.0, + "learning_rate": 1.4999999999999999e-05, + "loss": 9.5935, + "step": 17 + }, + { + "epoch": 1.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \")\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Hiding in\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 9.580382347106934, + "eval_runtime": 2.2821, + "eval_samples_per_second": 59.156, + "step": 17 + }, + { + "epoch": 2.0, + "learning_rate": 2.9999999999999997e-05, + "loss": 9.0928, + "step": 34 + }, + { + "epoch": 2.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \")\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"99.5%\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Hiding in\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 9.190422058105469, + "eval_runtime": 2.2408, + "eval_samples_per_second": 60.245, + "step": 34 + }, + { + "epoch": 3.0, + "learning_rate": 4.4999999999999996e-05, + "loss": 8.1515, + "step": 51 + }, + { + "epoch": 3.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0.0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification: climate\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 8.223907470703125, + "eval_runtime": 2.3068, + "eval_samples_per_second": 58.523, + "step": 51 + }, + { + "epoch": 4.0, + "learning_rate": 5.9999999999999995e-05, + "loss": 6.6637, + "step": 68 + }, + { + "epoch": 4.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0.0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Andrew Neil\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"please show us\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"very goi\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 6.206799030303955, + "eval_runtime": 2.267, + "eval_samples_per_second": 59.549, + "step": 68 + }, + { + "epoch": 5.0, + "learning_rate": 7.5e-05, + "loss": 4.5728, + "step": 85 + }, + { + "epoch": 5.0, + "eval_accuracy": 0.0, + "eval_average_metrics": 0.0, + "eval_classification_report": "{\"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"1\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"33000f\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Andrew Neil\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"CNN\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"COVID\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Florida\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"India\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"dal\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"doubtful\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"if de\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"microwave\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_d\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"troll\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"trud\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"u\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", + "eval_f1_macro": 0.0, + "eval_f1_micro": 0.0, + "eval_f1_weighted": 0.0, + "eval_loss": 3.01625657081604, + "eval_runtime": 2.3046, + "eval_samples_per_second": 58.579, + "step": 85 + }, + { + "epoch": 6.0, + "learning_rate": 8.999999999999999e-05, + "loss": 2.1563, + "step": 102 + }, + { + "epoch": 6.0, + "eval_accuracy": 25.925925925925924, + "eval_average_metrics": 18.478859129440288, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"challenge\": {\"precision\": 0.27049180327868855, \"recall\": 0.9166666666666666, \"f1-score\": 0.4177215189873418, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.125, \"f1-score\": 0.2222222222222222, \"support\": 16.0}, \"statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.25925925925925924, \"macro avg\": {\"precision\": 0.11549925484351714, \"recall\": 0.09469696969696968, \"f1-score\": 0.058176703746324, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.19064966605950215, \"recall\": 0.25925925925925924, \"f1-score\": 0.13772985362296192, \"support\": 135.0}}", + "eval_f1_macro": 7.99929676511955, + "eval_f1_micro": 26.217228464419474, + "eval_f1_weighted": 13.772985362296192, + "eval_loss": 1.1285314559936523, + "eval_runtime": 2.054, + "eval_samples_per_second": 65.725, + "step": 102 + }, + { + "epoch": 7.0, + "learning_rate": 0.00010499999999999999, + "loss": 1.0166, + "step": 119 + }, + { + "epoch": 7.0, + "eval_accuracy": 40.0, + "eval_average_metrics": 33.589515948427554, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.2727272727272727, \"f1-score\": 0.39999999999999997, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.16666666666666666, \"f1-score\": 0.26666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.3392857142857143, \"recall\": 0.9743589743589743, \"f1-score\": 0.5033112582781457, \"support\": 39.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.4375, \"f1-score\": 0.6086956521739131, \"support\": 16.0}, \"accuracy\": 0.4, \"macro avg\": {\"precision\": 0.3444940476190476, \"recall\": 0.23140661421911424, \"f1-score\": 0.22233419713984065, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.45542328042328045, \"recall\": 0.4, \"f1-score\": 0.3212464407972614, \"support\": 135.0}}", + "eval_f1_macro": 22.233419713984066, + "eval_f1_micro": 40.00000000000001, + "eval_f1_weighted": 32.12464407972614, + "eval_loss": 0.6725317239761353, + "eval_runtime": 2.2366, + "eval_samples_per_second": 60.359, + "step": 119 + }, + { + "epoch": 8.0, + "learning_rate": 0.00011999999999999999, + "loss": 0.7319, + "step": 136 + }, + { + "epoch": 8.0, + "eval_accuracy": 48.888888888888886, + "eval_average_metrics": 46.684741490249316, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.4, \"recall\": 0.2222222222222222, \"f1-score\": 0.2857142857142857, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.45454545454545453, \"f1-score\": 0.5555555555555556, \"support\": 11.0}, \"challenge\": {\"precision\": 0.4426229508196721, \"recall\": 0.75, \"f1-score\": 0.556701030927835, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.47619047619047616, \"recall\": 0.5128205128205128, \"f1-score\": 0.49382716049382713, \"support\": 39.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.25, \"recall\": 0.15384615384615385, \"f1-score\": 0.1904761904761905, \"support\": 13.0}, \"request\": {\"precision\": 0.8181818181818182, \"recall\": 0.5625, \"f1-score\": 0.6666666666666666, \"support\": 16.0}, \"accuracy\": 0.4888888888888889, \"macro avg\": {\"precision\": 0.5126601199347101, \"recall\": 0.39449179292929293, \"f1-score\": 0.4269509445626284, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.47632523517769415, \"recall\": 0.4888888888888889, \"f1-score\": 0.46266093726956653, \"support\": 135.0}}", + "eval_f1_macro": 42.69509445626284, + "eval_f1_micro": 48.888888888888886, + "eval_f1_weighted": 46.26609372695665, + "eval_loss": 0.5566144585609436, + "eval_runtime": 2.1877, + "eval_samples_per_second": 61.709, + "step": 136 + }, + { + "epoch": 9.0, + "learning_rate": 0.000135, + "loss": 0.6661, + "step": 153 + }, + { + "epoch": 9.0, + "eval_accuracy": 54.074074074074076, + "eval_average_metrics": 54.03755296584174, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.3333333333333333, \"f1-score\": 0.42857142857142855, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7777777777777778, \"recall\": 0.6363636363636364, \"f1-score\": 0.7000000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.5581395348837209, \"recall\": 0.6666666666666666, \"f1-score\": 0.6075949367088608, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.41025641025641024, \"f1-score\": 0.5245901639344261, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"rejection\": {\"precision\": 0.2727272727272727, \"recall\": 0.46153846153846156, \"f1-score\": 0.3428571428571428, \"support\": 13.0}, \"request\": {\"precision\": 0.4482758620689655, \"recall\": 0.8125, \"f1-score\": 0.5777777777777777, \"support\": 16.0}, \"accuracy\": 0.5407407407407407, \"macro avg\": {\"precision\": 0.6417741468413081, \"recall\": 0.5192489801864801, \"f1-score\": 0.5386995722568455, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.60651913713751, \"recall\": 0.5407407407407407, \"f1-score\": 0.5413210648953426, \"support\": 135.0}}", + "eval_f1_macro": 53.86995722568455, + "eval_f1_micro": 54.074074074074076, + "eval_f1_weighted": 54.13210648953427, + "eval_loss": 0.511962890625, + "eval_runtime": 2.1819, + "eval_samples_per_second": 61.874, + "step": 153 + }, + { + "epoch": 10.0, + "learning_rate": 0.00015, + "loss": 0.6131, + "step": 170 + }, + { + "epoch": 10.0, + "eval_accuracy": 57.03703703703704, + "eval_average_metrics": 57.547429298101335, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6111111111111112, \"recall\": 0.6111111111111112, \"f1-score\": 0.6111111111111112, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7391304347826086, \"recall\": 0.4358974358974359, \"f1-score\": 0.5483870967741935, \"support\": 39.0}, \"question\": {\"precision\": 0.8, \"recall\": 0.4444444444444444, \"f1-score\": 0.5714285714285714, \"support\": 9.0}, \"rejection\": {\"precision\": 0.3157894736842105, \"recall\": 0.46153846153846156, \"f1-score\": 0.37499999999999994, \"support\": 13.0}, \"request\": {\"precision\": 0.42424242424242425, \"recall\": 0.875, \"f1-score\": 0.5714285714285714, \"support\": 16.0}, \"accuracy\": 0.5703703703703704, \"macro avg\": {\"precision\": 0.6664789856723495, \"recall\": 0.576352466977467, \"f1-score\": 0.5870368430852302, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6322059283767903, \"recall\": 0.5703703703703704, \"f1-score\": 0.5741195880980826, \"support\": 135.0}}", + "eval_f1_macro": 58.703684308523016, + "eval_f1_micro": 57.03703703703704, + "eval_f1_weighted": 57.41195880980826, + "eval_loss": 0.4787319600582123, + "eval_runtime": 2.2206, + "eval_samples_per_second": 60.794, + "step": 170 + }, + { + "epoch": 11.0, + "learning_rate": 0.000165, + "loss": 0.5787, + "step": 187 + }, + { + "epoch": 11.0, + "eval_accuracy": 64.44444444444444, + "eval_average_metrics": 65.62118906429522, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6333333333333333, \"recall\": 0.5277777777777778, \"f1-score\": 0.5757575757575758, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.5833333333333334, \"recall\": 0.717948717948718, \"f1-score\": 0.6436781609195402, \"support\": 39.0}, \"question\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.5384615384615384, \"f1-score\": 0.608695652173913, \"support\": 13.0}, \"request\": {\"precision\": 0.7333333333333333, \"recall\": 0.6875, \"f1-score\": 0.7096774193548386, \"support\": 16.0}, \"accuracy\": 0.6444444444444445, \"macro avg\": {\"precision\": 0.7084821428571428, \"recall\": 0.6890115093240092, \"f1-score\": 0.6939908069080865, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6505908289241622, \"recall\": 0.6444444444444445, \"f1-score\": 0.6419678667748333, \"support\": 135.0}}", + "eval_f1_macro": 69.39908069080865, + "eval_f1_micro": 64.44444444444444, + "eval_f1_weighted": 64.19678667748333, + "eval_loss": 0.4315463900566101, + "eval_runtime": 2.336, + "eval_samples_per_second": 57.791, + "step": 187 + }, + { + "epoch": 12.0, + "learning_rate": 0.00017999999999999998, + "loss": 0.5394, + "step": 204 + }, + { + "epoch": 12.0, + "eval_accuracy": 68.14814814814815, + "eval_average_metrics": 69.08949458669183, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.8181818181818182, \"f1-score\": 0.6923076923076923, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6444444444444445, \"recall\": 0.8055555555555556, \"f1-score\": 0.7160493827160493, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6896551724137931, \"recall\": 0.5128205128205128, \"f1-score\": 0.5882352941176471, \"support\": 39.0}, \"question\": {\"precision\": 0.8333333333333334, \"recall\": 0.5555555555555556, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 13.0}, \"request\": {\"precision\": 0.7333333333333333, \"recall\": 0.6875, \"f1-score\": 0.7096774193548386, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7395188623636899, \"recall\": 0.7354858682983683, \"f1-score\": 0.7244826695435831, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6953210858957987, \"recall\": 0.6814814814814815, \"f1-score\": 0.6761341509611271, \"support\": 135.0}}", + "eval_f1_macro": 72.4482669543583, + "eval_f1_micro": 68.14814814814815, + "eval_f1_weighted": 67.6134150961127, + "eval_loss": 0.3929964601993561, + "eval_runtime": 2.2653, + "eval_samples_per_second": 59.595, + "step": 204 + }, + { + "epoch": 13.0, + "learning_rate": 0.000195, + "loss": 0.4445, + "step": 221 + }, + { + "epoch": 13.0, + "eval_accuracy": 68.14814814814815, + "eval_average_metrics": 69.54531108529271, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6136363636363636, \"recall\": 0.75, \"f1-score\": 0.6749999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.631578947368421, \"recall\": 0.6153846153846154, \"f1-score\": 0.6233766233766234, \"support\": 39.0}, \"question\": {\"precision\": 0.8571428571428571, \"recall\": 0.6666666666666666, \"f1-score\": 0.75, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8571428571428571, \"recall\": 0.46153846153846156, \"f1-score\": 0.6, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7767885896997739, \"recall\": 0.7221311674436675, \"f1-score\": 0.7386757543926662, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6968149297096665, \"recall\": 0.6814814814814815, \"f1-score\": 0.6801737260560791, \"support\": 135.0}}", + "eval_f1_macro": 73.86757543926662, + "eval_f1_micro": 68.14814814814815, + "eval_f1_weighted": 68.0173726056079, + "eval_loss": 0.3637169301509857, + "eval_runtime": 2.1262, + "eval_samples_per_second": 63.494, + "step": 221 + }, + { + "epoch": 14.0, + "learning_rate": 0.00020999999999999998, + "loss": 0.4284, + "step": 238 + }, + { + "epoch": 14.0, + "eval_accuracy": 68.88888888888889, + "eval_average_metrics": 70.04225470037278, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.6363636363636364, \"f1-score\": 0.6666666666666666, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6511627906976745, \"recall\": 0.7777777777777778, \"f1-score\": 0.7088607594936709, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5384615384615384, \"f1-score\": 0.6176470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 0.6923076923076923, \"f1-score\": 0.7826086956521738, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7628809968099262, \"recall\": 0.7316554972804973, \"f1-score\": 0.7363724073834267, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.711127376440127, \"recall\": 0.6888888888888889, \"f1-score\": 0.6875400028537069, \"support\": 135.0}}", + "eval_f1_macro": 73.63724073834267, + "eval_f1_micro": 68.88888888888889, + "eval_f1_weighted": 68.7540002853707, + "eval_loss": 0.35276439785957336, + "eval_runtime": 2.1486, + "eval_samples_per_second": 62.832, + "step": 238 + }, + { + "epoch": 15.0, + "learning_rate": 0.000225, + "loss": 0.4125, + "step": 255 + }, + { + "epoch": 15.0, + "eval_accuracy": 66.66666666666666, + "eval_average_metrics": 67.82947247449044, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.5918367346938775, \"recall\": 0.8055555555555556, \"f1-score\": 0.6823529411764706, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8571428571428571, \"recall\": 0.3076923076923077, \"f1-score\": 0.45283018867924535, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6153846153846154, \"f1-score\": 0.761904761904762, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.6666666666666666, \"macro avg\": {\"precision\": 0.761308929166072, \"recall\": 0.7724237567987569, \"f1-score\": 0.7343863238519761, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7304495152114199, \"recall\": 0.6666666666666666, \"f1-score\": 0.6454592417943081, \"support\": 135.0}}", + "eval_f1_macro": 73.43863238519761, + "eval_f1_micro": 66.66666666666666, + "eval_f1_weighted": 64.54592417943081, + "eval_loss": 0.36039280891418457, + "eval_runtime": 2.0664, + "eval_samples_per_second": 65.332, + "step": 255 + }, + { + "epoch": 16.0, + "learning_rate": 0.00023999999999999998, + "loss": 0.3537, + "step": 272 + }, + { + "epoch": 16.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.72736523652416, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6521739130434783, \"recall\": 0.8333333333333334, \"f1-score\": 0.7317073170731708, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8095238095238095, \"recall\": 0.4358974358974359, \"f1-score\": 0.5666666666666667, \"support\": 39.0}, \"question\": {\"precision\": 0.8333333333333334, \"recall\": 0.5555555555555556, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6923076923076923, \"f1-score\": 0.8181818181818181, \"support\": 13.0}, \"request\": {\"precision\": 0.5185185185185185, \"recall\": 0.875, \"f1-score\": 0.6511627906976744, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.8048186968023925, \"recall\": 0.7737592268842269, \"f1-score\": 0.7645378088268715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7534887832391858, \"recall\": 0.7037037037037037, \"f1-score\": 0.6971493932266877, \"support\": 135.0}}", + "eval_f1_macro": 76.45378088268716, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 69.71493932266877, + "eval_loss": 0.3415175974369049, + "eval_runtime": 2.2011, + "eval_samples_per_second": 61.334, + "step": 272 + }, + { + "epoch": 17.0, + "learning_rate": 0.00025499999999999996, + "loss": 0.3307, + "step": 289 + }, + { + "epoch": 17.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.80907924352218, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7720217737121258, \"recall\": 0.7883704836829837, \"f1-score\": 0.7742597755505364, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7391776099879956, \"recall\": 0.725925925925926, \"f1-score\": 0.7262515423384989, \"support\": 135.0}}", + "eval_f1_macro": 77.42597755505363, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.62515423384988, + "eval_loss": 0.320279061794281, + "eval_runtime": 2.1547, + "eval_samples_per_second": 62.654, + "step": 289 + }, + { + "epoch": 18.0, + "learning_rate": 0.00027, + "loss": 0.2829, + "step": 306 + }, + { + "epoch": 18.0, + "eval_accuracy": 75.55555555555556, + "eval_average_metrics": 76.66401588989308, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6829268292682927, \"recall\": 0.7777777777777778, \"f1-score\": 0.7272727272727273, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.7857142857142857, \"recall\": 0.6875, \"f1-score\": 0.7333333333333334, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.7981489039415868, \"recall\": 0.807868346930847, \"f1-score\": 0.7989500265816055, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7661238851482753, \"recall\": 0.7555555555555555, \"f1-score\": 0.7564994979030066, \"support\": 135.0}}", + "eval_f1_macro": 79.89500265816055, + "eval_f1_micro": 75.55555555555556, + "eval_f1_weighted": 75.64994979030067, + "eval_loss": 0.3107147812843323, + "eval_runtime": 2.1712, + "eval_samples_per_second": 62.178, + "step": 306 + }, + { + "epoch": 19.0, + "learning_rate": 0.000285, + "loss": 0.27, + "step": 323 + }, + { + "epoch": 19.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.65875801590278, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6078431372549019, \"recall\": 0.8611111111111112, \"f1-score\": 0.7126436781609194, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8823529411764706, \"recall\": 0.38461538461538464, \"f1-score\": 0.5357142857142858, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7684180606974724, \"recall\": 0.7708758255633257, \"f1-score\": 0.7515478083829743, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7415855386443622, \"recall\": 0.6962962962962963, \"f1-score\": 0.6822099196605441, \"support\": 135.0}}", + "eval_f1_macro": 75.15478083829744, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 68.22099196605443, + "eval_loss": 0.3932558298110962, + "eval_runtime": 2.1067, + "eval_samples_per_second": 64.082, + "step": 323 + }, + { + "epoch": 20.0, + "learning_rate": 0.0003, + "loss": 0.2814, + "step": 340 + }, + { + "epoch": 20.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.63929543961528, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6904761904761905, \"recall\": 0.7435897435897436, \"f1-score\": 0.7160493827160495, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7865632631257631, \"recall\": 0.731479458041958, \"f1-score\": 0.7439023969088441, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7466218966218966, \"recall\": 0.7333333333333333, \"f1-score\": 0.7350027540091005, \"support\": 135.0}}", + "eval_f1_macro": 74.39023969088441, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.50027540091006, + "eval_loss": 0.35566049814224243, + "eval_runtime": 2.1665, + "eval_samples_per_second": 62.312, + "step": 340 + }, + { + "epoch": 21.0, + "learning_rate": 0.00029833333333333334, + "loss": 0.242, + "step": 357 + }, + { + "epoch": 21.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.95524042950514, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7941800524037366, \"recall\": 0.8099504662004662, \"f1-score\": 0.7964169000933707, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7463545226703122, \"recall\": 0.7333333333333333, \"f1-score\": 0.7351260504201681, \"support\": 135.0}}", + "eval_f1_macro": 79.64169000933707, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.51260504201682, + "eval_loss": 0.33881324529647827, + "eval_runtime": 2.177, + "eval_samples_per_second": 62.012, + "step": 357 + }, + { + "epoch": 22.0, + "learning_rate": 0.00029666666666666665, + "loss": 0.1889, + "step": 374 + }, + { + "epoch": 22.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.62607689125785, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6511627906976745, \"recall\": 0.7777777777777778, \"f1-score\": 0.7088607594936709, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7855620155038759, \"recall\": 0.8003350815850816, \"f1-score\": 0.7871122658545654, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7467298306057996, \"recall\": 0.7333333333333333, \"f1-score\": 0.7312641431290826, \"support\": 135.0}}", + "eval_f1_macro": 78.71122658545654, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.12641431290825, + "eval_loss": 0.38223153352737427, + "eval_runtime": 2.1885, + "eval_samples_per_second": 61.685, + "step": 374 + }, + { + "epoch": 23.0, + "learning_rate": 0.00029499999999999996, + "loss": 0.1787, + "step": 391 + }, + { + "epoch": 23.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.05134386837818, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5384615384615384, \"f1-score\": 0.6176470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7613135503335141, \"recall\": 0.7901065947940948, \"f1-score\": 0.7691766030736619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7284590496326734, \"recall\": 0.7185185185185186, \"f1-score\": 0.7158401146244283, \"support\": 135.0}}", + "eval_f1_macro": 76.91766030736619, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 71.58401146244283, + "eval_loss": 0.40378537774086, + "eval_runtime": 2.2888, + "eval_samples_per_second": 58.982, + "step": 391 + }, + { + "epoch": 24.0, + "learning_rate": 0.00029333333333333327, + "loss": 0.1458, + "step": 408 + }, + { + "epoch": 24.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.88129852922421, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.5128205128205128, \"f1-score\": 0.625, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.9230769230769231, \"f1-score\": 0.8275862068965517, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7539396367521367, \"recall\": 0.8122207653457654, \"f1-score\": 0.7701591825594307, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7255460588793923, \"recall\": 0.7037037037037037, \"f1-score\": 0.6976853512021299, \"support\": 135.0}}", + "eval_f1_macro": 77.01591825594308, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 69.768535120213, + "eval_loss": 0.4484809637069702, + "eval_runtime": 2.2039, + "eval_samples_per_second": 61.256, + "step": 408 + }, + { + "epoch": 25.0, + "learning_rate": 0.00029166666666666664, + "loss": 0.1755, + "step": 425 + }, + { + "epoch": 25.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.30296325475025, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6595744680851063, \"recall\": 0.8611111111111112, \"f1-score\": 0.7469879518072289, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8461538461538461, \"recall\": 0.5641025641025641, \"f1-score\": 0.676923076923077, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7791251301889599, \"recall\": 0.7255244755244756, \"f1-score\": 0.7344306944480657, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7577047066408767, \"recall\": 0.7333333333333333, \"f1-score\": 0.7310211690752781, \"support\": 135.0}}", + "eval_f1_macro": 73.44306944480657, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.1021169075278, + "eval_loss": 0.47192898392677307, + "eval_runtime": 2.1604, + "eval_samples_per_second": 62.489, + "step": 425 + }, + { + "epoch": 26.0, + "learning_rate": 0.00029, + "loss": 0.1193, + "step": 442 + }, + { + "epoch": 26.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.94665416513004, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6578947368421053, \"recall\": 0.6944444444444444, \"f1-score\": 0.6756756756756757, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8037195260879472, \"recall\": 0.7995337995337994, \"f1-score\": 0.7956619448395763, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7480504291030607, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355375550989586, \"support\": 135.0}}", + "eval_f1_macro": 79.56619448395764, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.55375550989585, + "eval_loss": 0.45392414927482605, + "eval_runtime": 2.2435, + "eval_samples_per_second": 60.174, + "step": 442 + }, + { + "epoch": 27.0, + "learning_rate": 0.0002883333333333333, + "loss": 0.1263, + "step": 459 + }, + { + "epoch": 27.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 74.0950271183071, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65, \"recall\": 0.7222222222222222, \"f1-score\": 0.6842105263157895, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7858405483405484, \"recall\": 0.7965957653457654, \"f1-score\": 0.7843223575106921, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7482972582972582, \"recall\": 0.725925925925926, \"f1-score\": 0.7276268753697399, \"support\": 135.0}}", + "eval_f1_macro": 78.43223575106921, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.762687536974, + "eval_loss": 0.46841171383857727, + "eval_runtime": 2.1169, + "eval_samples_per_second": 63.773, + "step": 459 + }, + { + "epoch": 28.0, + "learning_rate": 0.0002866666666666667, + "loss": 0.1043, + "step": 476 + }, + { + "epoch": 28.0, + "eval_accuracy": 68.88888888888889, + "eval_average_metrics": 69.42866207163016, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8333333333333334, \"recall\": 0.5128205128205128, \"f1-score\": 0.6349206349206349, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7394724025974025, \"recall\": 0.7357651029526029, \"f1-score\": 0.7130452134668899, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7225299021965689, \"recall\": 0.6888888888888889, \"f1-score\": 0.6863234916205385, \"support\": 135.0}}", + "eval_f1_macro": 71.30452134668899, + "eval_f1_micro": 68.88888888888889, + "eval_f1_weighted": 68.63234916205386, + "eval_loss": 0.49356982111930847, + "eval_runtime": 2.0925, + "eval_samples_per_second": 64.518, + "step": 476 + }, + { + "epoch": 29.0, + "learning_rate": 0.000285, + "loss": 0.1174, + "step": 493 + }, + { + "epoch": 29.0, + "eval_accuracy": 68.14814814814815, + "eval_average_metrics": 69.89790682417316, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6206896551724138, \"recall\": 0.5, \"f1-score\": 0.5538461538461539, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.4827586206896552, \"recall\": 0.875, \"f1-score\": 0.6222222222222222, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7515477604986782, \"recall\": 0.7737592268842268, \"f1-score\": 0.7502473425267543, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7103981308931253, \"recall\": 0.6814814814814815, \"f1-score\": 0.6827059674772094, \"support\": 135.0}}", + "eval_f1_macro": 75.02473425267543, + "eval_f1_micro": 68.14814814814815, + "eval_f1_weighted": 68.27059674772093, + "eval_loss": 0.5587795972824097, + "eval_runtime": 2.2187, + "eval_samples_per_second": 60.847, + "step": 493 + }, + { + "epoch": 30.0, + "learning_rate": 0.0002833333333333333, + "loss": 0.0689, + "step": 510 + }, + { + "epoch": 30.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.94861619800345, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.5897435897435898, \"f1-score\": 0.676470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7848776232528335, \"recall\": 0.7861790986790986, \"f1-score\": 0.7791600854100853, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7445433448587359, \"recall\": 0.725925925925926, \"f1-score\": 0.7269327106582008, \"support\": 135.0}}", + "eval_f1_macro": 77.91600854100852, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.69327106582007, + "eval_loss": 0.539673388004303, + "eval_runtime": 2.1316, + "eval_samples_per_second": 63.333, + "step": 510 + }, + { + "epoch": 31.0, + "learning_rate": 0.00028166666666666666, + "loss": 0.0539, + "step": 527 + }, + { + "epoch": 31.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.33781120741372, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5641025641025641, \"f1-score\": 0.6285714285714286, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7561203917050692, \"recall\": 0.7387031371406372, \"f1-score\": 0.7282430580928023, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7082807077430734, \"recall\": 0.6962962962962963, \"f1-score\": 0.6926767976111539, \"support\": 135.0}}", + "eval_f1_macro": 72.82430580928023, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.26767976111539, + "eval_loss": 0.5683187246322632, + "eval_runtime": 2.1767, + "eval_samples_per_second": 62.021, + "step": 527 + }, + { + "epoch": 32.0, + "learning_rate": 0.00028, + "loss": 0.0591, + "step": 544 + }, + { + "epoch": 32.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.58032751159402, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7027027027027027, \"recall\": 0.7222222222222222, \"f1-score\": 0.7123287671232876, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7916666666666666, \"recall\": 0.48717948717948717, \"f1-score\": 0.6031746031746031, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.738858691983692, \"recall\": 0.7976641414141414, \"f1-score\": 0.757487853810853, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7226086476086476, \"recall\": 0.7037037037037037, \"f1-score\": 0.6983178392455008, \"support\": 135.0}}", + "eval_f1_macro": 75.7487853810853, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 69.83178392455008, + "eval_loss": 0.6107626557350159, + "eval_runtime": 2.1901, + "eval_samples_per_second": 61.641, + "step": 544 + }, + { + "epoch": 33.0, + "learning_rate": 0.00027833333333333334, + "loss": 0.0539, + "step": 561 + }, + { + "epoch": 33.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.67097529451648, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7804387991887991, \"recall\": 0.7399718337218337, \"f1-score\": 0.7468834223239258, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7408524975191642, \"recall\": 0.7333333333333333, \"f1-score\": 0.7332889227900669, \"support\": 135.0}}", + "eval_f1_macro": 74.68834223239257, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.32889227900668, + "eval_loss": 0.5886625051498413, + "eval_runtime": 2.2726, + "eval_samples_per_second": 59.404, + "step": 561 + }, + { + "epoch": 34.0, + "learning_rate": 0.00027666666666666665, + "loss": 0.0734, + "step": 578 + }, + { + "epoch": 34.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 75.0892478093565, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7929446284709443, \"recall\": 0.7491744366744366, \"f1-score\": 0.757480864546082, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7589443694706852, \"recall\": 0.7481481481481481, \"f1-score\": 0.749792751531882, \"support\": 135.0}}", + "eval_f1_macro": 75.7480864546082, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 74.9792751531882, + "eval_loss": 0.5679999589920044, + "eval_runtime": 2.2089, + "eval_samples_per_second": 61.116, + "step": 578 + }, + { + "epoch": 35.0, + "learning_rate": 0.00027499999999999996, + "loss": 0.0343, + "step": 595 + }, + { + "epoch": 35.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.3823414203499, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7803482087018672, \"recall\": 0.7343628593628594, \"f1-score\": 0.7409159272416515, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7359272164150212, \"recall\": 0.7185185185185186, \"f1-score\": 0.7173406925353076, \"support\": 135.0}}", + "eval_f1_macro": 74.09159272416514, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 71.73406925353078, + "eval_loss": 0.64279705286026, + "eval_runtime": 2.2258, + "eval_samples_per_second": 60.654, + "step": 595 + }, + { + "epoch": 36.0, + "learning_rate": 0.00027333333333333333, + "loss": 0.0256, + "step": 612 + }, + { + "epoch": 36.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.27192027743499, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.6111111111111112, \"f1-score\": 0.6285714285714287, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.774467344387318, \"recall\": 0.7373008935508936, \"f1-score\": 0.7363179712444419, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7299169542548802, \"recall\": 0.7037037037037037, \"f1-score\": 0.7071514324455502, \"support\": 135.0}}", + "eval_f1_macro": 73.63179712444419, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.71514324455502, + "eval_loss": 0.6609192490577698, + "eval_runtime": 2.2472, + "eval_samples_per_second": 60.074, + "step": 612 + }, + { + "epoch": 37.0, + "learning_rate": 0.00027166666666666664, + "loss": 0.0369, + "step": 629 + }, + { + "epoch": 37.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.77168024471725, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7861613248710023, \"recall\": 0.8058651418026419, \"f1-score\": 0.7918380042511899, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7395884688357807, \"recall\": 0.7333333333333333, \"f1-score\": 0.7323625388708334, \"support\": 135.0}}", + "eval_f1_macro": 79.18380042511899, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.23625388708334, + "eval_loss": 0.6587929725646973, + "eval_runtime": 2.1685, + "eval_samples_per_second": 62.254, + "step": 629 + }, + { + "epoch": 38.0, + "learning_rate": 0.00027, + "loss": 0.0411, + "step": 646 + }, + { + "epoch": 38.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.4984910288592, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7916610299422799, \"recall\": 0.7314915986790986, \"f1-score\": 0.741379475703325, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7429092512425847, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215231284140066, \"support\": 135.0}}", + "eval_f1_macro": 74.13794757033249, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.15231284140066, + "eval_loss": 0.717117190361023, + "eval_runtime": 2.2579, + "eval_samples_per_second": 59.789, + "step": 646 + }, + { + "epoch": 39.0, + "learning_rate": 0.0002683333333333333, + "loss": 0.0188, + "step": 663 + }, + { + "epoch": 39.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 75.21675953993977, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7982278836574502, \"recall\": 0.7576000388500388, \"f1-score\": 0.7635135515393372, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7587398134869756, \"recall\": 0.7481481481481481, \"f1-score\": 0.7488605337619573, \"support\": 135.0}}", + "eval_f1_macro": 76.35135515393372, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 74.88605337619573, + "eval_loss": 0.6625694036483765, + "eval_runtime": 2.1219, + "eval_samples_per_second": 63.622, + "step": 663 + }, + { + "epoch": 40.0, + "learning_rate": 0.0002666666666666666, + "loss": 0.0232, + "step": 680 + }, + { + "epoch": 40.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.41849073937567, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7912581597759617, \"recall\": 0.8052641802641802, \"f1-score\": 0.7945949432404541, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7460020576119648, \"recall\": 0.7407407407407407, \"f1-score\": 0.7406632048530915, \"support\": 135.0}}", + "eval_f1_macro": 79.45949432404542, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.06632048530913, + "eval_loss": 0.7407746315002441, + "eval_runtime": 2.1843, + "eval_samples_per_second": 61.805, + "step": 680 + }, + { + "epoch": 41.0, + "learning_rate": 0.000265, + "loss": 0.034, + "step": 697 + }, + { + "epoch": 41.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 72.10801080339358, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6285714285714286, \"recall\": 0.6111111111111112, \"f1-score\": 0.619718309859155, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7691468253968254, \"recall\": 0.7905193764568765, \"f1-score\": 0.7703084495012319, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327513227513227, \"recall\": 0.7037037037037037, \"f1-score\": 0.7066045752271038, \"support\": 135.0}}", + "eval_f1_macro": 77.0308449501232, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.66045752271039, + "eval_loss": 0.8014240264892578, + "eval_runtime": 2.1576, + "eval_samples_per_second": 62.569, + "step": 697 + }, + { + "epoch": 42.0, + "learning_rate": 0.0002633333333333333, + "loss": 0.0393, + "step": 714 + }, + { + "epoch": 42.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 72.91137239666652, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7874594155844157, \"recall\": 0.7123154623154623, \"f1-score\": 0.7361968097262215, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.741950456950457, \"recall\": 0.725925925925926, \"f1-score\": 0.7284062342885872, \"support\": 135.0}}", + "eval_f1_macro": 73.61968097262215, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.84062342885873, + "eval_loss": 0.7516341805458069, + "eval_runtime": 2.2389, + "eval_samples_per_second": 60.297, + "step": 714 + }, + { + "epoch": 43.0, + "learning_rate": 0.00026166666666666667, + "loss": 0.0187, + "step": 731 + }, + { + "epoch": 43.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 71.6853552122398, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.5833333333333334, \"f1-score\": 0.6086956521739131, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.5897435897435898, \"f1-score\": 0.6571428571428573, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.784555908547844, \"recall\": 0.7870471542346542, \"f1-score\": 0.7754307414417895, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7231971245710792, \"recall\": 0.6962962962962963, \"f1-score\": 0.6993908744552095, \"support\": 135.0}}", + "eval_f1_macro": 77.54307414417895, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.93908744552097, + "eval_loss": 0.8274635672569275, + "eval_runtime": 2.2726, + "eval_samples_per_second": 59.402, + "step": 731 + }, + { + "epoch": 44.0, + "learning_rate": 0.00026, + "loss": 0.0326, + "step": 748 + }, + { + "epoch": 44.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.29350182491167, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.778793688949939, \"recall\": 0.7780205905205906, \"f1-score\": 0.7740684627673118, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7306166056166056, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206345731921179, \"support\": 135.0}}", + "eval_f1_macro": 77.40684627673117, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06345731921179, + "eval_loss": 0.7537038326263428, + "eval_runtime": 2.251, + "eval_samples_per_second": 59.972, + "step": 748 + }, + { + "epoch": 45.0, + "learning_rate": 0.00025833333333333334, + "loss": 0.0446, + "step": 765 + }, + { + "epoch": 45.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.55936282460405, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7762371615312792, \"recall\": 0.7913752913752914, \"f1-score\": 0.7800986983223825, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7445440398381575, \"recall\": 0.7333333333333333, \"f1-score\": 0.735609147995113, \"support\": 135.0}}", + "eval_f1_macro": 78.00986983223825, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.5609147995113, + "eval_loss": 0.7628147006034851, + "eval_runtime": 2.1831, + "eval_samples_per_second": 61.838, + "step": 765 + }, + { + "epoch": 46.0, + "learning_rate": 0.00025666666666666665, + "loss": 0.0222, + "step": 782 + }, + { + "epoch": 46.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.61743270582245, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7865956727518593, \"recall\": 0.8073462995337995, \"f1-score\": 0.7881557617778407, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7388550822627902, \"recall\": 0.7185185185185186, \"f1-score\": 0.71950450941802, \"support\": 135.0}}", + "eval_f1_macro": 78.81557617778408, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 71.950450941802, + "eval_loss": 0.8409448266029358, + "eval_runtime": 2.2893, + "eval_samples_per_second": 58.969, + "step": 782 + }, + { + "epoch": 47.0, + "learning_rate": 0.00025499999999999996, + "loss": 0.0137, + "step": 799 + }, + { + "epoch": 47.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 72.17480498243586, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8064516129032258, \"recall\": 0.6410256410256411, \"f1-score\": 0.7142857142857142, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7435458256022771, \"recall\": 0.7664687742812744, \"f1-score\": 0.7519798629432732, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7228986803180352, \"recall\": 0.7111111111111111, \"f1-score\": 0.712790114131939, \"support\": 135.0}}", + "eval_f1_macro": 75.19798629432732, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.27901141319391, + "eval_loss": 0.8047698140144348, + "eval_runtime": 2.1643, + "eval_samples_per_second": 62.375, + "step": 799 + }, + { + "epoch": 48.0, + "learning_rate": 0.00025333333333333333, + "loss": 0.0209, + "step": 816 + }, + { + "epoch": 48.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.5841574785611, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.48, \"recall\": 0.75, \"f1-score\": 0.5853658536585366, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.78078081232493, \"recall\": 0.8030060217560218, \"f1-score\": 0.7854538367301481, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7372613341632949, \"recall\": 0.7185185185185186, \"f1-score\": 0.7208754253752583, \"support\": 135.0}}", + "eval_f1_macro": 78.54538367301481, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.08754253752583, + "eval_loss": 0.8688657283782959, + "eval_runtime": 2.1847, + "eval_samples_per_second": 61.794, + "step": 816 + }, + { + "epoch": 49.0, + "learning_rate": 0.00025166666666666664, + "loss": 0.0091, + "step": 833 + }, + { + "epoch": 49.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.75127888430008, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8125, \"recall\": 0.6666666666666666, \"f1-score\": 0.7323943661971831, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7851934523809524, \"recall\": 0.7992667055167055, \"f1-score\": 0.7875581374336444, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.748179012345679, \"recall\": 0.7333333333333333, \"f1-score\": 0.735826351271692, \"support\": 135.0}}", + "eval_f1_macro": 78.75581374336443, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.58263512716921, + "eval_loss": 0.8018043637275696, + "eval_runtime": 2.2936, + "eval_samples_per_second": 58.859, + "step": 833 + }, + { + "epoch": 50.0, + "learning_rate": 0.00025, + "loss": 0.0097, + "step": 850 + }, + { + "epoch": 50.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.57898288245715, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65, \"recall\": 0.7222222222222222, \"f1-score\": 0.6842105263157895, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8387096774193549, \"recall\": 0.6666666666666666, \"f1-score\": 0.7428571428571428, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8063538611925709, \"recall\": 0.7948475135975136, \"f1-score\": 0.795720639514387, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7636070381231672, \"recall\": 0.7407407407407407, \"f1-score\": 0.7459571943024179, \"support\": 135.0}}", + "eval_f1_macro": 79.5720639514387, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.59571943024179, + "eval_loss": 0.8517288565635681, + "eval_runtime": 2.1756, + "eval_samples_per_second": 62.051, + "step": 850 + }, + { + "epoch": 51.0, + "learning_rate": 0.0002483333333333333, + "loss": 0.0193, + "step": 867 + }, + { + "epoch": 51.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 72.8732174399158, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6153846153846154, \"f1-score\": 0.6575342465753424, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7702113042186571, \"recall\": 0.8102175602175602, \"f1-score\": 0.7806177067478437, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7336914900640391, \"recall\": 0.7111111111111111, \"f1-score\": 0.7120887686265667, \"support\": 135.0}}", + "eval_f1_macro": 78.06177067478437, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.20887686265667, + "eval_loss": 0.9536928534507751, + "eval_runtime": 2.1272, + "eval_samples_per_second": 63.463, + "step": 867 + }, + { + "epoch": 52.0, + "learning_rate": 0.0002466666666666666, + "loss": 0.0046, + "step": 884 + }, + { + "epoch": 52.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.64170538878341, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6829268292682927, \"recall\": 0.717948717948718, \"f1-score\": 0.7000000000000001, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7706033088556692, \"recall\": 0.7170563811188811, \"f1-score\": 0.7315823058619944, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7196237044807725, \"recall\": 0.7111111111111111, \"f1-score\": 0.7118636874671196, \"support\": 135.0}}", + "eval_f1_macro": 73.15823058619944, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.18636874671196, + "eval_loss": 0.8713467121124268, + "eval_runtime": 2.2641, + "eval_samples_per_second": 59.627, + "step": 884 + }, + { + "epoch": 53.0, + "learning_rate": 0.000245, + "loss": 0.0085, + "step": 901 + }, + { + "epoch": 53.0, + "eval_accuracy": 68.88888888888889, + "eval_average_metrics": 69.60698215446021, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.5277777777777778, \"f1-score\": 0.6031746031746033, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7484503734503735, \"recall\": 0.721930846930847, \"f1-score\": 0.7174509189640768, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7075614075614075, \"recall\": 0.6888888888888889, \"f1-score\": 0.6890505894365544, \"support\": 135.0}}", + "eval_f1_macro": 71.74509189640769, + "eval_f1_micro": 68.88888888888889, + "eval_f1_weighted": 68.90505894365543, + "eval_loss": 0.9699578285217285, + "eval_runtime": 2.2097, + "eval_samples_per_second": 61.095, + "step": 901 + }, + { + "epoch": 54.0, + "learning_rate": 0.0002433333333333333, + "loss": 0.0181, + "step": 918 + }, + { + "epoch": 54.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.51024557847745, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.6363636363636364, \"f1-score\": 0.6666666666666666, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.717948717948718, \"f1-score\": 0.708860759493671, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7724702380952382, \"recall\": 0.7056138306138307, \"f1-score\": 0.7247131858217115, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7247113997113998, \"recall\": 0.7111111111111111, \"f1-score\": 0.7134744150951646, \"support\": 135.0}}", + "eval_f1_macro": 72.47131858217115, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.34744150951646, + "eval_loss": 0.9296635389328003, + "eval_runtime": 2.1343, + "eval_samples_per_second": 63.252, + "step": 918 + }, + { + "epoch": 55.0, + "learning_rate": 0.00024166666666666664, + "loss": 0.0055, + "step": 935 + }, + { + "epoch": 55.0, + "eval_accuracy": 68.14814814814815, + "eval_average_metrics": 69.88210719785789, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6451612903225806, \"recall\": 0.5555555555555556, \"f1-score\": 0.5970149253731343, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.7857142857142857, \"recall\": 0.8461538461538461, \"f1-score\": 0.8148148148148148, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7426706358157971, \"recall\": 0.7828950563325563, \"f1-score\": 0.751917200243583, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7054854941951716, \"recall\": 0.6814814814814815, \"f1-score\": 0.6804041247077698, \"support\": 135.0}}", + "eval_f1_macro": 75.1917200243583, + "eval_f1_micro": 68.14814814814815, + "eval_f1_weighted": 68.04041247077699, + "eval_loss": 0.9763553142547607, + "eval_runtime": 2.1762, + "eval_samples_per_second": 62.035, + "step": 935 + }, + { + "epoch": 56.0, + "learning_rate": 0.00023999999999999998, + "loss": 0.0051, + "step": 952 + }, + { + "epoch": 56.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.26569664902999, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7671443391552086, \"recall\": 0.7989206973581974, \"f1-score\": 0.7732298136645963, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7413002188364506, \"recall\": 0.7185185185185186, \"f1-score\": 0.720361015259566, \"support\": 135.0}}", + "eval_f1_macro": 77.32298136645963, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.0361015259566, + "eval_loss": 0.9423373341560364, + "eval_runtime": 2.238, + "eval_samples_per_second": 60.322, + "step": 952 + }, + { + "epoch": 57.0, + "learning_rate": 0.0002383333333333333, + "loss": 0.0114, + "step": 969 + }, + { + "epoch": 57.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 72.53354760332923, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.6363636363636364, \"f1-score\": 0.6363636363636364, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7597564519879099, \"recall\": 0.7908835955710956, \"f1-score\": 0.7683601409092514, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7236771810173345, \"recall\": 0.7111111111111111, \"f1-score\": 0.7107595410016957, \"support\": 135.0}}", + "eval_f1_macro": 76.83601409092513, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.07595410016957, + "eval_loss": 0.9203009605407715, + "eval_runtime": 2.1386, + "eval_samples_per_second": 63.126, + "step": 969 + }, + { + "epoch": 58.0, + "learning_rate": 0.00023666666666666663, + "loss": 0.0072, + "step": 986 + }, + { + "epoch": 58.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.6642490769342, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"cancellation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7071727257211129, \"recall\": 0.7029213070879737, \"f1-score\": 0.6958756590822236, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7471616316777607, \"recall\": 0.7185185185185186, \"f1-score\": 0.7240017370132812, \"support\": 135.0}}", + "eval_f1_macro": 78.28601164675014, + "eval_f1_micro": 72.1189591078067, + "eval_f1_weighted": 72.40017370132811, + "eval_loss": 0.9572186470031738, + "eval_runtime": 2.1451, + "eval_samples_per_second": 62.935, + "step": 986 + }, + { + "epoch": 59.0, + "learning_rate": 0.00023499999999999997, + "loss": 0.0042, + "step": 1003 + }, + { + "epoch": 59.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.7649842113928, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6153846153846154, \"f1-score\": 0.6486486486486486, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7765705163233214, \"recall\": 0.7346967268842268, \"f1-score\": 0.736597489423982, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7294317448174391, \"recall\": 0.7111111111111111, \"f1-score\": 0.7117796568095075, \"support\": 135.0}}", + "eval_f1_macro": 73.6597489423982, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.17796568095075, + "eval_loss": 0.8926369547843933, + "eval_runtime": 2.1674, + "eval_samples_per_second": 62.286, + "step": 1003 + }, + { + "epoch": 60.0, + "learning_rate": 0.0002333333333333333, + "loss": 0.0108, + "step": 1020 + }, + { + "epoch": 60.0, + "eval_accuracy": 67.4074074074074, + "eval_average_metrics": 67.96942582571397, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.5555555555555556, \"f1-score\": 0.6153846153846154, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6923076923076923, \"f1-score\": 0.8181818181818181, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.674074074074074, \"macro avg\": {\"precision\": 0.7355337802536078, \"recall\": 0.6994160353535354, \"f1-score\": 0.6963058347068645, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6961066081755738, \"recall\": 0.674074074074074, \"f1-score\": 0.674323050173546, \"support\": 135.0}}", + "eval_f1_macro": 69.63058347068645, + "eval_f1_micro": 67.4074074074074, + "eval_f1_weighted": 67.4323050173546, + "eval_loss": 0.9399789571762085, + "eval_runtime": 2.1159, + "eval_samples_per_second": 63.802, + "step": 1020 + }, + { + "epoch": 61.0, + "learning_rate": 0.00023166666666666667, + "loss": 0.0056, + "step": 1037 + }, + { + "epoch": 61.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 71.17152096147001, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.5277777777777778, \"f1-score\": 0.6031746031746033, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7443091168091167, \"recall\": 0.7997219794094794, \"f1-score\": 0.7591938053549183, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.719369420702754, \"recall\": 0.6962962962962963, \"f1-score\": 0.6950744405112893, \"support\": 135.0}}", + "eval_f1_macro": 75.91938053549183, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.50744405112893, + "eval_loss": 0.9959737062454224, + "eval_runtime": 2.1462, + "eval_samples_per_second": 62.902, + "step": 1037 + }, + { + "epoch": 62.0, + "learning_rate": 0.00023, + "loss": 0.0091, + "step": 1054 + }, + { + "epoch": 62.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.43021846765396, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7951003737641669, \"recall\": 0.7485856157731158, \"f1-score\": 0.754229181551163, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7622480329376882, \"recall\": 0.7407407407407407, \"f1-score\": 0.7414980756735141, \"support\": 135.0}}", + "eval_f1_macro": 75.42291815511629, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.14980756735142, + "eval_loss": 0.9423182010650635, + "eval_runtime": 2.0725, + "eval_samples_per_second": 65.137, + "step": 1054 + }, + { + "epoch": 63.0, + "learning_rate": 0.0002283333333333333, + "loss": 0.006, + "step": 1071 + }, + { + "epoch": 63.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.13630195174072, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.5833333333333334, \"f1-score\": 0.626865671641791, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7705405145929339, \"recall\": 0.7309574106449107, \"f1-score\": 0.7339954289925987, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7191460431245379, \"recall\": 0.7037037037037037, \"f1-score\": 0.7040492416696227, \"support\": 135.0}}", + "eval_f1_macro": 73.39954289925987, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.40492416696227, + "eval_loss": 0.9775393009185791, + "eval_runtime": 2.1426, + "eval_samples_per_second": 63.009, + "step": 1071 + }, + { + "epoch": 64.0, + "learning_rate": 0.00022666666666666663, + "loss": 0.0071, + "step": 1088 + }, + { + "epoch": 64.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.50660628560995, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8333333333333334, \"recall\": 0.6410256410256411, \"f1-score\": 0.7246376811594204, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5185185185185185, \"recall\": 0.875, \"f1-score\": 0.6511627906976744, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7855794668294669, \"recall\": 0.7343507187257188, \"f1-score\": 0.7410246999652906, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7480406698925217, \"recall\": 0.7185185185185186, \"f1-score\": 0.7222025144220701, \"support\": 135.0}}", + "eval_f1_macro": 74.10246999652907, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.22025144220703, + "eval_loss": 1.0484141111373901, + "eval_runtime": 2.2145, + "eval_samples_per_second": 60.962, + "step": 1088 + }, + { + "epoch": 65.0, + "learning_rate": 0.000225, + "loss": 0.0034, + "step": 1105 + }, + { + "epoch": 65.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.9244200133918, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.675, \"recall\": 0.6923076923076923, \"f1-score\": 0.6835443037974683, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5789473684210527, \"recall\": 0.6875, \"f1-score\": 0.6285714285714286, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7900560761394801, \"recall\": 0.7485734751359752, \"f1-score\": 0.7557596878082226, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7442556314967181, \"recall\": 0.7333333333333333, \"f1-score\": 0.7345504460607825, \"support\": 135.0}}", + "eval_f1_macro": 75.57596878082227, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.45504460607825, + "eval_loss": 0.943881094455719, + "eval_runtime": 2.2664, + "eval_samples_per_second": 59.565, + "step": 1105 + }, + { + "epoch": 66.0, + "learning_rate": 0.00022333333333333333, + "loss": 0.007, + "step": 1122 + }, + { + "epoch": 66.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.91255932928726, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65625, \"recall\": 0.5833333333333334, \"f1-score\": 0.6176470588235293, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7927684294871795, \"recall\": 0.7234787781662781, \"f1-score\": 0.7393785014005602, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7358262108262108, \"recall\": 0.7111111111111111, \"f1-score\": 0.7149016495487084, \"support\": 135.0}}", + "eval_f1_macro": 73.93785014005603, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.49016495487082, + "eval_loss": 1.0599663257598877, + "eval_runtime": 2.1786, + "eval_samples_per_second": 61.965, + "step": 1122 + }, + { + "epoch": 67.0, + "learning_rate": 0.00022166666666666667, + "loss": 0.0042, + "step": 1139 + }, + { + "epoch": 67.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 75.12632131391055, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8064516129032258, \"recall\": 0.6410256410256411, \"f1-score\": 0.7142857142857142, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7984916367980883, \"recall\": 0.8247074106449106, \"f1-score\": 0.801042885612959, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7643603699278729, \"recall\": 0.7333333333333333, \"f1-score\": 0.7373433002767967, \"support\": 135.0}}", + "eval_f1_macro": 80.1042885612959, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.73433002767968, + "eval_loss": 1.0870040655136108, + "eval_runtime": 2.253, + "eval_samples_per_second": 59.919, + "step": 1139 + }, + { + "epoch": 68.0, + "learning_rate": 0.00021999999999999995, + "loss": 0.0048, + "step": 1156 + }, + { + "epoch": 68.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.55212011915016, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.791516354016354, \"recall\": 0.7966625388500388, \"f1-score\": 0.7820397286821705, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7515169329984145, \"recall\": 0.7185185185185186, \"f1-score\": 0.7230080390467987, \"support\": 135.0}}", + "eval_f1_macro": 78.20397286821705, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.30080390467987, + "eval_loss": 1.0407289266586304, + "eval_runtime": 2.2127, + "eval_samples_per_second": 61.011, + "step": 1156 + }, + { + "epoch": 69.0, + "learning_rate": 0.0002183333333333333, + "loss": 0.0026, + "step": 1173 + }, + { + "epoch": 69.0, + "eval_accuracy": 68.88888888888889, + "eval_average_metrics": 70.6530531645798, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5, \"f1-score\": 0.5901639344262295, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7635660173160174, \"recall\": 0.783040743978244, \"f1-score\": 0.7556388197313891, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.731313772647106, \"recall\": 0.6888888888888889, \"f1-score\": 0.6927055290740252, \"support\": 135.0}}", + "eval_f1_macro": 75.56388197313892, + "eval_f1_micro": 68.88888888888889, + "eval_f1_weighted": 69.27055290740252, + "eval_loss": 1.0904052257537842, + "eval_runtime": 2.1381, + "eval_samples_per_second": 63.141, + "step": 1173 + }, + { + "epoch": 70.0, + "learning_rate": 0.00021666666666666666, + "loss": 0.0014, + "step": 1190 + }, + { + "epoch": 70.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.87504553531681, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7893565188172043, \"recall\": 0.814290743978244, \"f1-score\": 0.7918166361805148, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7587290919952211, \"recall\": 0.7333333333333333, \"f1-score\": 0.7365185185654917, \"support\": 135.0}}", + "eval_f1_macro": 79.18166361805147, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.65185185654917, + "eval_loss": 1.0420516729354858, + "eval_runtime": 2.151, + "eval_samples_per_second": 62.763, + "step": 1190 + }, + { + "epoch": 71.0, + "learning_rate": 0.000215, + "loss": 0.0047, + "step": 1207 + }, + { + "epoch": 71.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.96275924849695, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7944225059094784, \"recall\": 0.7552629662004662, \"f1-score\": 0.7562933597057457, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7554414742767257, \"recall\": 0.7333333333333333, \"f1-score\": 0.735550343567466, \"support\": 135.0}}", + "eval_f1_macro": 75.62933597057457, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.5550343567466, + "eval_loss": 1.0715558528900146, + "eval_runtime": 2.2105, + "eval_samples_per_second": 61.073, + "step": 1207 + }, + { + "epoch": 72.0, + "learning_rate": 0.00021333333333333333, + "loss": 0.0019, + "step": 1224 + }, + { + "epoch": 72.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.36672605045894, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7666666666666667, \"recall\": 0.6388888888888888, \"f1-score\": 0.696969696969697, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7698383520074696, \"recall\": 0.7404271076146076, \"f1-score\": 0.7363487615884876, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7427004876024484, \"recall\": 0.7185185185185186, \"f1-score\": 0.7212832433928325, \"support\": 135.0}}", + "eval_f1_macro": 73.63487615884875, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.12832433928324, + "eval_loss": 1.1170759201049805, + "eval_runtime": 2.1545, + "eval_samples_per_second": 62.659, + "step": 1224 + }, + { + "epoch": 73.0, + "learning_rate": 0.00021166666666666667, + "loss": 0.0009, + "step": 1241 + }, + { + "epoch": 73.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.45112685839807, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7725357975357976, \"recall\": 0.7965168512043512, \"f1-score\": 0.7786937025399319, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7492250958917624, \"recall\": 0.7333333333333333, \"f1-score\": 0.7326847051293242, \"support\": 135.0}}", + "eval_f1_macro": 77.8693702539932, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.26847051293241, + "eval_loss": 1.0454473495483398, + "eval_runtime": 2.0945, + "eval_samples_per_second": 64.453, + "step": 1241 + }, + { + "epoch": 74.0, + "learning_rate": 0.00020999999999999998, + "loss": 0.0137, + "step": 1258 + }, + { + "epoch": 74.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.70350070614597, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.6363636363636364, \"f1-score\": 0.6363636363636364, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6097560975609756, \"recall\": 0.6944444444444444, \"f1-score\": 0.6493506493506495, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7840743537137566, \"recall\": 0.7153627622377623, \"f1-score\": 0.7315785053858942, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7402746783789677, \"recall\": 0.7111111111111111, \"f1-score\": 0.7143393006377224, \"support\": 135.0}}", + "eval_f1_macro": 73.15785053858941, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.43393006377225, + "eval_loss": 1.0791348218917847, + "eval_runtime": 2.1939, + "eval_samples_per_second": 61.535, + "step": 1258 + }, + { + "epoch": 75.0, + "learning_rate": 0.00020833333333333332, + "loss": 0.0254, + "step": 1275 + }, + { + "epoch": 75.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.53066318197898, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7937821165762342, \"recall\": 0.7505766802641802, \"f1-score\": 0.7575329088486984, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7541625694566871, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422121369489791, \"support\": 135.0}}", + "eval_f1_macro": 75.75329088486984, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.22121369489791, + "eval_loss": 0.9704490303993225, + "eval_runtime": 2.225, + "eval_samples_per_second": 60.674, + "step": 1275 + }, + { + "epoch": 76.0, + "learning_rate": 0.00020666666666666666, + "loss": 0.0039, + "step": 1292 + }, + { + "epoch": 76.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.88592559677284, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7785046518667209, \"recall\": 0.7366877913752914, \"f1-score\": 0.7404468801773039, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7311833743557882, \"recall\": 0.7111111111111111, \"f1-score\": 0.7127679214713875, \"support\": 135.0}}", + "eval_f1_macro": 74.04468801773038, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.27679214713875, + "eval_loss": 1.000522255897522, + "eval_runtime": 2.2253, + "eval_samples_per_second": 60.667, + "step": 1292 + }, + { + "epoch": 77.0, + "learning_rate": 0.000205, + "loss": 0.0095, + "step": 1309 + }, + { + "epoch": 77.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.50938740714625, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7787457160894661, \"recall\": 0.7404271076146076, \"f1-score\": 0.7417628359660171, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7426006493506493, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215756232827957, \"support\": 135.0}}", + "eval_f1_macro": 74.17628359660172, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.15756232827957, + "eval_loss": 0.9987503886222839, + "eval_runtime": 2.0847, + "eval_samples_per_second": 64.758, + "step": 1309 + }, + { + "epoch": 78.0, + "learning_rate": 0.00020333333333333333, + "loss": 0.0016, + "step": 1326 + }, + { + "epoch": 78.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.50535308482088, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7787042370711725, \"recall\": 0.7404271076146076, \"f1-score\": 0.74266339591752, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7378141889969847, \"recall\": 0.7185185185185186, \"f1-score\": 0.7205136904382781, \"support\": 135.0}}", + "eval_f1_macro": 74.266339591752, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.0513690438278, + "eval_loss": 1.0020543336868286, + "eval_runtime": 2.1954, + "eval_samples_per_second": 61.492, + "step": 1326 + }, + { + "epoch": 79.0, + "learning_rate": 0.00020166666666666667, + "loss": 0.0011, + "step": 1343 + }, + { + "epoch": 79.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.48732962316897, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8027706290849673, \"recall\": 0.751790743978244, \"f1-score\": 0.7583791127061329, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7516668482207698, \"recall\": 0.725925925925926, \"f1-score\": 0.7292622203687736, \"support\": 135.0}}", + "eval_f1_macro": 75.8379112706133, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.92622203687736, + "eval_loss": 1.041650652885437, + "eval_runtime": 2.0975, + "eval_samples_per_second": 64.363, + "step": 1343 + }, + { + "epoch": 80.0, + "learning_rate": 0.00019999999999999998, + "loss": 0.0022, + "step": 1360 + }, + { + "epoch": 80.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.38039917163168, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7955693164362518, \"recall\": 0.751790743978244, \"f1-score\": 0.7553531897837845, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7471616316777607, \"recall\": 0.725925925925926, \"f1-score\": 0.7280109252296305, \"support\": 135.0}}", + "eval_f1_macro": 75.53531897837846, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.80109252296305, + "eval_loss": 1.0494800806045532, + "eval_runtime": 2.2088, + "eval_samples_per_second": 61.119, + "step": 1360 + }, + { + "epoch": 81.0, + "learning_rate": 0.00019833333333333332, + "loss": 0.0126, + "step": 1377 + }, + { + "epoch": 81.0, + "eval_accuracy": 68.88888888888889, + "eval_average_metrics": 69.48276901714948, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5, \"f1-score\": 0.5901639344262295, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7443162393162392, \"recall\": 0.7262711247086246, \"f1-score\": 0.7127222322360148, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.716459639126306, \"recall\": 0.6888888888888889, \"f1-score\": 0.6888107506721864, \"support\": 135.0}}", + "eval_f1_macro": 71.27222322360149, + "eval_f1_micro": 68.88888888888889, + "eval_f1_weighted": 68.88107506721865, + "eval_loss": 1.0959280729293823, + "eval_runtime": 2.2009, + "eval_samples_per_second": 61.34, + "step": 1377 + }, + { + "epoch": 82.0, + "learning_rate": 0.00019666666666666666, + "loss": 0.0033, + "step": 1394 + }, + { + "epoch": 82.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.0222652316685, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.717948717948718, \"f1-score\": 0.7567567567567569, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8054761904761905, \"recall\": 0.738678855866356, \"f1-score\": 0.7563368980085744, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7553862433862435, \"recall\": 0.7333333333333333, \"f1-score\": 0.737887044591499, \"support\": 135.0}}", + "eval_f1_macro": 75.63368980085744, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.78870445914991, + "eval_loss": 1.07533860206604, + "eval_runtime": 2.33, + "eval_samples_per_second": 57.939, + "step": 1394 + }, + { + "epoch": 83.0, + "learning_rate": 0.000195, + "loss": 0.0029, + "step": 1411 + }, + { + "epoch": 83.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.76246669403774, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6842105263157895, \"recall\": 0.7222222222222222, \"f1-score\": 0.7027027027027027, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8101666927736006, \"recall\": 0.7543949106449106, \"f1-score\": 0.7658868817992679, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7598260698041399, \"recall\": 0.7407407407407407, \"f1-score\": 0.7431303044807601, \"support\": 135.0}}", + "eval_f1_macro": 76.58868817992679, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.313030448076, + "eval_loss": 1.025146722793579, + "eval_runtime": 2.189, + "eval_samples_per_second": 61.672, + "step": 1411 + }, + { + "epoch": 84.0, + "learning_rate": 0.00019333333333333333, + "loss": 0.0008, + "step": 1428 + }, + { + "epoch": 84.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.65410439323483, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.766517611157791, \"recall\": 0.7387031371406372, \"f1-score\": 0.7312475747258356, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7362255787043392, \"recall\": 0.7111111111111111, \"f1-score\": 0.7126943787813353, \"support\": 135.0}}", + "eval_f1_macro": 73.12475747258355, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.26943787813353, + "eval_loss": 1.0640238523483276, + "eval_runtime": 2.1727, + "eval_samples_per_second": 62.136, + "step": 1428 + }, + { + "epoch": 85.0, + "learning_rate": 0.00019166666666666665, + "loss": 0.0007, + "step": 1445 + }, + { + "epoch": 85.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.63498328794381, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7612767620310724, \"recall\": 0.7369548853923854, \"f1-score\": 0.730588001969581, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7325891775317063, \"recall\": 0.7111111111111111, \"f1-score\": 0.7125891073259495, \"support\": 135.0}}", + "eval_f1_macro": 73.05880019695809, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.25891073259494, + "eval_loss": 1.0694376230239868, + "eval_runtime": 2.1511, + "eval_samples_per_second": 62.759, + "step": 1445 + }, + { + "epoch": 86.0, + "learning_rate": 0.00018999999999999998, + "loss": 0.0013, + "step": 1462 + }, + { + "epoch": 86.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.88851983050755, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6944444444444444, \"recall\": 0.6944444444444444, \"f1-score\": 0.6944444444444444, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7967806461352657, \"recall\": 0.7448462995337995, \"f1-score\": 0.7532179628803489, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.754411567364466, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356561636732861, \"support\": 135.0}}", + "eval_f1_macro": 75.3217962880349, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.56561636732862, + "eval_loss": 1.076446771621704, + "eval_runtime": 2.2436, + "eval_samples_per_second": 60.171, + "step": 1462 + }, + { + "epoch": 87.0, + "learning_rate": 0.00018833333333333332, + "loss": 0.0052, + "step": 1479 + }, + { + "epoch": 87.0, + "eval_accuracy": 77.03703703703704, + "eval_average_metrics": 77.53590008051296, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7941176470588235, \"recall\": 0.6923076923076923, \"f1-score\": 0.7397260273972601, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7703703703703704, \"macro avg\": {\"precision\": 0.833837516113058, \"recall\": 0.7720898892773893, \"f1-score\": 0.7884724218577965, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.784561747926041, \"recall\": 0.7703703703703704, \"f1-score\": 0.7722228406219811, \"support\": 135.0}}", + "eval_f1_macro": 78.84724218577965, + "eval_f1_micro": 77.03703703703704, + "eval_f1_weighted": 77.22228406219811, + "eval_loss": 1.0223504304885864, + "eval_runtime": 2.1607, + "eval_samples_per_second": 62.479, + "step": 1479 + }, + { + "epoch": 88.0, + "learning_rate": 0.00018666666666666666, + "loss": 0.0004, + "step": 1496 + }, + { + "epoch": 88.0, + "eval_accuracy": 75.55555555555556, + "eval_average_metrics": 76.05592523834689, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8130099067599068, \"recall\": 0.7651454448329449, \"f1-score\": 0.773336418091264, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7716006216006216, \"recall\": 0.7555555555555555, \"f1-score\": 0.7577894803315004, \"support\": 135.0}}", + "eval_f1_macro": 77.33364180912639, + "eval_f1_micro": 75.55555555555556, + "eval_f1_weighted": 75.77894803315003, + "eval_loss": 1.0519832372665405, + "eval_runtime": 2.2144, + "eval_samples_per_second": 60.966, + "step": 1496 + }, + { + "epoch": 89.0, + "learning_rate": 0.000185, + "loss": 0.0004, + "step": 1513 + }, + { + "epoch": 89.0, + "eval_accuracy": 75.55555555555556, + "eval_average_metrics": 76.11648469654651, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7714285714285715, \"recall\": 0.6923076923076923, \"f1-score\": 0.7297297297297296, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8263431013431014, \"recall\": 0.7546620046620047, \"f1-score\": 0.7756144827692142, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7708913308913309, \"recall\": 0.7555555555555555, \"f1-score\": 0.7579337939815347, \"support\": 135.0}}", + "eval_f1_macro": 77.56144827692142, + "eval_f1_micro": 75.55555555555556, + "eval_f1_weighted": 75.79337939815348, + "eval_loss": 1.0824300050735474, + "eval_runtime": 2.2212, + "eval_samples_per_second": 60.777, + "step": 1513 + }, + { + "epoch": 90.0, + "learning_rate": 0.00018333333333333334, + "loss": 0.0023, + "step": 1530 + }, + { + "epoch": 90.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89906381584439, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7837301587301587, \"recall\": 0.7384360431235432, \"f1-score\": 0.7390882467117281, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7427689594356262, \"recall\": 0.7111111111111111, \"f1-score\": 0.7146520836998245, \"support\": 135.0}}", + "eval_f1_macro": 73.90882467117281, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.46520836998246, + "eval_loss": 1.1680539846420288, + "eval_runtime": 2.2118, + "eval_samples_per_second": 61.037, + "step": 1530 + }, + { + "epoch": 91.0, + "learning_rate": 0.00018166666666666665, + "loss": 0.0017, + "step": 1547 + }, + { + "epoch": 91.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.54780705462385, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7821969696969697, \"recall\": 0.8073462995337995, \"f1-score\": 0.7829353295805501, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7491021324354658, \"recall\": 0.7185185185185186, \"f1-score\": 0.7219399155673665, \"support\": 135.0}}", + "eval_f1_macro": 78.29353295805501, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.19399155673665, + "eval_loss": 1.1076750755310059, + "eval_runtime": 2.1138, + "eval_samples_per_second": 63.867, + "step": 1547 + }, + { + "epoch": 92.0, + "learning_rate": 0.00017999999999999998, + "loss": 0.0072, + "step": 1564 + }, + { + "epoch": 92.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.88573213660933, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7721699167819858, \"recall\": 0.7942344114219114, \"f1-score\": 0.7758369725875447, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.742619934459015, \"recall\": 0.725925925925926, \"f1-score\": 0.7277404610249767, \"support\": 135.0}}", + "eval_f1_macro": 77.58369725875447, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.77404610249766, + "eval_loss": 1.0486587285995483, + "eval_runtime": 2.2394, + "eval_samples_per_second": 60.284, + "step": 1564 + }, + { + "epoch": 93.0, + "learning_rate": 0.00017833333333333332, + "loss": 0.0057, + "step": 1581 + }, + { + "epoch": 93.0, + "eval_accuracy": 68.88888888888889, + "eval_average_metrics": 69.75310065197965, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.5, \"f1-score\": 0.5806451612903226, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7741794389820706, \"recall\": 0.7117023601398601, \"f1-score\": 0.7183544685246277, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7288036330141594, \"recall\": 0.6888888888888889, \"f1-score\": 0.693991779776781, \"support\": 135.0}}", + "eval_f1_macro": 71.83544685246277, + "eval_f1_micro": 68.88888888888889, + "eval_f1_weighted": 69.3991779776781, + "eval_loss": 1.1878422498703003, + "eval_runtime": 2.1986, + "eval_samples_per_second": 61.403, + "step": 1581 + }, + { + "epoch": 94.0, + "learning_rate": 0.00017666666666666666, + "loss": 0.0195, + "step": 1598 + }, + { + "epoch": 94.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.61304930742271, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7979249011857708, \"recall\": 0.7587351884226884, \"f1-score\": 0.7599361706778587, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7623871565900551, \"recall\": 0.7407407407407407, \"f1-score\": 0.7431043201375682, \"support\": 135.0}}", + "eval_f1_macro": 75.99361706778586, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.31043201375682, + "eval_loss": 1.0754647254943848, + "eval_runtime": 2.2267, + "eval_samples_per_second": 60.629, + "step": 1598 + }, + { + "epoch": 95.0, + "learning_rate": 0.000175, + "loss": 0.0009, + "step": 1615 + }, + { + "epoch": 95.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.42176515283398, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7921437115516063, \"recall\": 0.7429523601398601, \"f1-score\": 0.7544827980154067, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7502160282862037, \"recall\": 0.7407407407407407, \"f1-score\": 0.7409063266164715, \"support\": 135.0}}", + "eval_f1_macro": 75.44827980154068, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.09063266164715, + "eval_loss": 1.0278376340866089, + "eval_runtime": 2.155, + "eval_samples_per_second": 62.646, + "step": 1615 + }, + { + "epoch": 96.0, + "learning_rate": 0.0001733333333333333, + "loss": 0.0004, + "step": 1632 + }, + { + "epoch": 96.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 75.17056142900891, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7297297297297297, \"recall\": 0.75, \"f1-score\": 0.7397260273972601, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7977492664992665, \"recall\": 0.7543159965034965, \"f1-score\": 0.7625258134847175, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7574262074262075, \"recall\": 0.7481481481481481, \"f1-score\": 0.7480003473793427, \"support\": 135.0}}", + "eval_f1_macro": 76.25258134847175, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 74.80003473793427, + "eval_loss": 1.0315518379211426, + "eval_runtime": 2.137, + "eval_samples_per_second": 63.172, + "step": 1632 + }, + { + "epoch": 97.0, + "learning_rate": 0.00017166666666666665, + "loss": 0.0004, + "step": 1649 + }, + { + "epoch": 97.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 75.17542277410698, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7647058823529411, \"recall\": 0.7222222222222222, \"f1-score\": 0.7428571428571428, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7985009658384271, \"recall\": 0.7494415306915307, \"f1-score\": 0.7611168071694387, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7574858211390718, \"recall\": 0.7481481481481481, \"f1-score\": 0.7496038074985443, \"support\": 135.0}}", + "eval_f1_macro": 76.11168071694388, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 74.96038074985442, + "eval_loss": 1.0185447931289673, + "eval_runtime": 2.2095, + "eval_samples_per_second": 61.099, + "step": 1649 + }, + { + "epoch": 98.0, + "learning_rate": 0.00016999999999999999, + "loss": 0.0081, + "step": 1666 + }, + { + "epoch": 98.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.67129308878316, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7769345152528264, \"recall\": 0.8063993298368298, \"f1-score\": 0.784722770729386, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7505437952307024, \"recall\": 0.7333333333333333, \"f1-score\": 0.735462286155274, \"support\": 135.0}}", + "eval_f1_macro": 78.4722770729386, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.5462286155274, + "eval_loss": 1.0540401935577393, + "eval_runtime": 2.1385, + "eval_samples_per_second": 63.127, + "step": 1666 + }, + { + "epoch": 99.0, + "learning_rate": 0.00016833333333333332, + "loss": 0.0003, + "step": 1683 + }, + { + "epoch": 99.0, + "eval_accuracy": 75.55555555555556, + "eval_average_metrics": 76.05696734786595, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7428571428571429, \"recall\": 0.7222222222222222, \"f1-score\": 0.732394366197183, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8140947288006112, \"recall\": 0.7654125388500388, \"f1-score\": 0.7737856961026234, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7716894434541492, \"recall\": 0.7555555555555555, \"f1-score\": 0.7573818867009036, \"support\": 135.0}}", + "eval_f1_macro": 77.37856961026233, + "eval_f1_micro": 75.55555555555556, + "eval_f1_weighted": 75.73818867009035, + "eval_loss": 1.054269790649414, + "eval_runtime": 2.1572, + "eval_samples_per_second": 62.58, + "step": 1683 + }, + { + "epoch": 100.0, + "learning_rate": 0.00016666666666666666, + "loss": 0.0016, + "step": 1700 + }, + { + "epoch": 100.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 75.19184257143311, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7990435981204232, \"recall\": 0.7535147144522145, \"f1-score\": 0.7606214838883731, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7652767059388452, \"recall\": 0.7481481481481481, \"f1-score\": 0.7507559226726549, \"support\": 135.0}}", + "eval_f1_macro": 76.06214838883731, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 75.07559226726548, + "eval_loss": 1.0762172937393188, + "eval_runtime": 2.2054, + "eval_samples_per_second": 61.213, + "step": 1700 + }, + { + "epoch": 101.0, + "learning_rate": 0.000165, + "loss": 0.0008, + "step": 1717 + }, + { + "epoch": 101.0, + "eval_accuracy": 76.29629629629629, + "eval_average_metrics": 76.48385951719736, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8235294117647058, \"recall\": 0.717948717948718, \"f1-score\": 0.767123287671233, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.762962962962963, \"macro avg\": {\"precision\": 0.8140522875816993, \"recall\": 0.748227466977467, \"f1-score\": 0.7677001436238999, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7779084967320261, \"recall\": 0.762962962962963, \"f1-score\": 0.7657283111380684, \"support\": 135.0}}", + "eval_f1_macro": 76.77001436238999, + "eval_f1_micro": 76.29629629629629, + "eval_f1_weighted": 76.57283111380684, + "eval_loss": 1.0158424377441406, + "eval_runtime": 2.1602, + "eval_samples_per_second": 62.494, + "step": 1717 + }, + { + "epoch": 102.0, + "learning_rate": 0.0001633333333333333, + "loss": 0.0005, + "step": 1734 + }, + { + "epoch": 102.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 72.98289629063926, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7834636668306023, \"recall\": 0.7290634712509713, \"f1-score\": 0.7370998050821901, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.752118603301399, \"recall\": 0.725925925925926, \"f1-score\": 0.7303641946915284, \"support\": 135.0}}", + "eval_f1_macro": 73.709980508219, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 73.03641946915283, + "eval_loss": 1.0503590106964111, + "eval_runtime": 2.2165, + "eval_samples_per_second": 60.907, + "step": 1734 + }, + { + "epoch": 103.0, + "learning_rate": 0.00016166666666666665, + "loss": 0.0013, + "step": 1751 + }, + { + "epoch": 103.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.67005589980079, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8069940476190476, \"recall\": 0.752324932012432, \"f1-score\": 0.7613581176647959, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7638364975619877, \"recall\": 0.7407407407407407, \"f1-score\": 0.7439626368457539, \"support\": 135.0}}", + "eval_f1_macro": 76.13581176647959, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.39626368457539, + "eval_loss": 1.0846141576766968, + "eval_runtime": 2.2097, + "eval_samples_per_second": 61.093, + "step": 1751 + }, + { + "epoch": 104.0, + "learning_rate": 0.00015999999999999999, + "loss": 0.0017, + "step": 1768 + }, + { + "epoch": 104.0, + "eval_accuracy": 67.4074074074074, + "eval_average_metrics": 67.85678034161822, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.65, \"recall\": 0.6666666666666666, \"f1-score\": 0.6582278481012659, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6153846153846154, \"f1-score\": 0.761904761904762, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.6875, \"f1-score\": 0.5789473684210527, \"support\": 16.0}, \"accuracy\": 0.674074074074074, \"macro avg\": {\"precision\": 0.7351551226551227, \"recall\": 0.6922166375291375, \"f1-score\": 0.689884997303385, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7053102453102452, \"recall\": 0.674074074074074, \"f1-score\": 0.6762380682131959, \"support\": 135.0}}", + "eval_f1_macro": 68.9884997303385, + "eval_f1_micro": 67.4074074074074, + "eval_f1_weighted": 67.62380682131959, + "eval_loss": 1.1709874868392944, + "eval_runtime": 2.145, + "eval_samples_per_second": 62.936, + "step": 1768 + }, + { + "epoch": 105.0, + "learning_rate": 0.00015833333333333332, + "loss": 0.0039, + "step": 1785 + }, + { + "epoch": 105.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.57626910983969, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6410256410256411, \"recall\": 0.6410256410256411, \"f1-score\": 0.6410256410256411, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.6875, \"f1-score\": 0.5789473684210527, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7753345792138895, \"recall\": 0.7230781371406372, \"f1-score\": 0.7304419012808433, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7212663028754984, \"recall\": 0.6962962962962963, \"f1-score\": 0.7000162705201517, \"support\": 135.0}}", + "eval_f1_macro": 73.04419012808432, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 70.00162705201517, + "eval_loss": 1.1890654563903809, + "eval_runtime": 2.2179, + "eval_samples_per_second": 60.869, + "step": 1785 + }, + { + "epoch": 106.0, + "learning_rate": 0.00015666666666666666, + "loss": 0.0005, + "step": 1802 + }, + { + "epoch": 106.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.60340419650365, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7788739038739039, \"recall\": 0.7317586926961928, \"f1-score\": 0.7311720214719142, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7352949519616188, \"recall\": 0.6962962962962963, \"f1-score\": 0.7003715537956388, \"support\": 135.0}}", + "eval_f1_macro": 73.11720214719142, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 70.03715537956387, + "eval_loss": 1.2529244422912598, + "eval_runtime": 2.2315, + "eval_samples_per_second": 60.498, + "step": 1802 + }, + { + "epoch": 107.0, + "learning_rate": 0.000155, + "loss": 0.001, + "step": 1819 + }, + { + "epoch": 107.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.58287100496284, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7766354016354016, \"recall\": 0.7317586926961928, \"f1-score\": 0.7314957452929186, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7284128765610247, \"recall\": 0.6962962962962963, \"f1-score\": 0.6992265023130029, \"support\": 135.0}}", + "eval_f1_macro": 73.14957452929185, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.92265023130028, + "eval_loss": 1.2114461660385132, + "eval_runtime": 2.1003, + "eval_samples_per_second": 64.278, + "step": 1819 + }, + { + "epoch": 108.0, + "learning_rate": 0.0001533333333333333, + "loss": 0.0002, + "step": 1836 + }, + { + "epoch": 108.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.58367232228676, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.77743083993084, \"recall\": 0.7317586926961928, \"f1-score\": 0.7323127643359169, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7265083265083264, \"recall\": 0.6962962962962963, \"f1-score\": 0.698441535962961, \"support\": 135.0}}", + "eval_f1_macro": 73.23127643359169, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.8441535962961, + "eval_loss": 1.21000075340271, + "eval_runtime": 2.1341, + "eval_samples_per_second": 63.257, + "step": 1836 + }, + { + "epoch": 109.0, + "learning_rate": 0.00015166666666666665, + "loss": 0.0041, + "step": 1853 + }, + { + "epoch": 109.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.92307044896559, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7983946608946609, \"recall\": 0.744779526029526, \"f1-score\": 0.7552466149013464, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7484559884559885, \"recall\": 0.7333333333333333, \"f1-score\": 0.7350095363906106, \"support\": 135.0}}", + "eval_f1_macro": 75.52466149013463, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.50095363906107, + "eval_loss": 1.1299362182617188, + "eval_runtime": 2.2527, + "eval_samples_per_second": 59.928, + "step": 1853 + }, + { + "epoch": 110.0, + "learning_rate": 0.00015, + "loss": 0.0012, + "step": 1870 + }, + { + "epoch": 110.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.93367080496068, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.798580771374889, \"recall\": 0.744512432012432, \"f1-score\": 0.7553214077392971, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7485956238897415, \"recall\": 0.7333333333333333, \"f1-score\": 0.7353587577924635, \"support\": 135.0}}", + "eval_f1_macro": 75.53214077392971, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.53587577924635, + "eval_loss": 1.1259392499923706, + "eval_runtime": 2.1641, + "eval_samples_per_second": 62.382, + "step": 1870 + }, + { + "epoch": 111.0, + "learning_rate": 0.00014833333333333332, + "loss": 0.0016, + "step": 1887 + }, + { + "epoch": 111.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.5803126086232, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7890165755297334, \"recall\": 0.7332277097902098, \"f1-score\": 0.7449370393743795, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7351884762411078, \"recall\": 0.7185185185185186, \"f1-score\": 0.7212384279335117, \"support\": 135.0}}", + "eval_f1_macro": 74.49370393743794, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.12384279335117, + "eval_loss": 1.1060606241226196, + "eval_runtime": 2.1888, + "eval_samples_per_second": 61.678, + "step": 1887 + }, + { + "epoch": 112.0, + "learning_rate": 0.00014666666666666664, + "loss": 0.0008, + "step": 1904 + }, + { + "epoch": 112.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.58815358410081, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6578947368421053, \"recall\": 0.6410256410256411, \"f1-score\": 0.6493506493506495, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.776365629984051, \"recall\": 0.727418414918415, \"f1-score\": 0.7315299716534531, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7238407466477642, \"recall\": 0.6962962962962963, \"f1-score\": 0.6994035791179866, \"support\": 135.0}}", + "eval_f1_macro": 73.15299716534531, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.94035791179867, + "eval_loss": 1.204077959060669, + "eval_runtime": 2.1515, + "eval_samples_per_second": 62.746, + "step": 1904 + }, + { + "epoch": 113.0, + "learning_rate": 0.000145, + "loss": 0.0009, + "step": 1921 + }, + { + "epoch": 113.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.56981914255991, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7873426359180036, \"recall\": 0.7421753593628594, \"f1-score\": 0.7447551523701981, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7428055060408002, \"recall\": 0.7185185185185186, \"f1-score\": 0.721000576295161, \"support\": 135.0}}", + "eval_f1_macro": 74.4755152370198, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.1000576295161, + "eval_loss": 1.2206599712371826, + "eval_runtime": 2.1932, + "eval_samples_per_second": 61.553, + "step": 1921 + }, + { + "epoch": 114.0, + "learning_rate": 0.00014333333333333334, + "loss": 0.0008, + "step": 1938 + }, + { + "epoch": 114.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.50895922645417, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7836120335230867, \"recall\": 0.7421753593628594, \"f1-score\": 0.7418505576965078, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7454529482043714, \"recall\": 0.7185185185185186, \"f1-score\": 0.7214707743246217, \"support\": 135.0}}", + "eval_f1_macro": 74.18505576965077, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.14707743246217, + "eval_loss": 1.201111078262329, + "eval_runtime": 2.1314, + "eval_samples_per_second": 63.338, + "step": 1938 + }, + { + "epoch": 115.0, + "learning_rate": 0.00014166666666666665, + "loss": 0.0003, + "step": 1955 + }, + { + "epoch": 115.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89831172580645, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7827099116161615, \"recall\": 0.7343628593628594, \"f1-score\": 0.7401694157684197, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7314534231200898, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135408310416161, \"support\": 135.0}}", + "eval_f1_macro": 74.01694157684197, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.35408310416163, + "eval_loss": 1.181020736694336, + "eval_runtime": 2.237, + "eval_samples_per_second": 60.349, + "step": 1955 + }, + { + "epoch": 116.0, + "learning_rate": 0.00014, + "loss": 0.0002, + "step": 1972 + }, + { + "epoch": 116.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.25460880133727, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6578947368421053, \"recall\": 0.6410256410256411, \"f1-score\": 0.6493506493506495, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7801606074952954, \"recall\": 0.7308906371406372, \"f1-score\": 0.7360371831919146, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7287160867554882, \"recall\": 0.7037037037037037, \"f1-score\": 0.7067397614541691, \"support\": 135.0}}", + "eval_f1_macro": 73.60371831919146, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.6739761454169, + "eval_loss": 1.216983675956726, + "eval_runtime": 2.1268, + "eval_samples_per_second": 63.474, + "step": 1972 + }, + { + "epoch": 117.0, + "learning_rate": 0.00013833333333333333, + "loss": 0.0015, + "step": 1989 + }, + { + "epoch": 117.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.36956241306592, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8001194760365956, \"recall\": 0.7453804875679876, \"f1-score\": 0.7551997801408894, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7443039001734654, \"recall\": 0.725925925925926, \"f1-score\": 0.7277308645298954, \"support\": 135.0}}", + "eval_f1_macro": 75.51997801408893, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.77308645298953, + "eval_loss": 1.1754764318466187, + "eval_runtime": 2.3156, + "eval_samples_per_second": 58.299, + "step": 1989 + }, + { + "epoch": 118.0, + "learning_rate": 0.00013666666666666666, + "loss": 0.0011, + "step": 2006 + }, + { + "epoch": 118.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.17485160829882, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7760871136872282, \"recall\": 0.7306235431235432, \"f1-score\": 0.7325914102148916, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7313728393819928, \"recall\": 0.7037037037037037, \"f1-score\": 0.7069952467096542, \"support\": 135.0}}", + "eval_f1_macro": 73.25914102148916, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.69952467096542, + "eval_loss": 1.1750222444534302, + "eval_runtime": 2.2858, + "eval_samples_per_second": 59.061, + "step": 2006 + }, + { + "epoch": 119.0, + "learning_rate": 0.000135, + "loss": 0.0001, + "step": 2023 + }, + { + "epoch": 119.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.73002733490209, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7730584332146831, \"recall\": 0.7262043512043512, \"f1-score\": 0.7334226043582956, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7285210035210035, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135562668155659, \"support\": 135.0}}", + "eval_f1_macro": 73.34226043582956, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.35562668155659, + "eval_loss": 1.1494803428649902, + "eval_runtime": 2.2653, + "eval_samples_per_second": 59.593, + "step": 2023 + }, + { + "epoch": 120.0, + "learning_rate": 0.0001333333333333333, + "loss": 0.0004, + "step": 2040 + }, + { + "epoch": 120.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.61782293955142, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8034016636957814, \"recall\": 0.752324932012432, \"f1-score\": 0.7609662631209947, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564920917862095, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422651729795805, \"support\": 135.0}}", + "eval_f1_macro": 76.09662631209946, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.22651729795805, + "eval_loss": 1.1421889066696167, + "eval_runtime": 2.1875, + "eval_samples_per_second": 61.713, + "step": 2040 + }, + { + "epoch": 121.0, + "learning_rate": 0.00013166666666666665, + "loss": 0.0023, + "step": 2057 + }, + { + "epoch": 121.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.5217489806628, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.5897435897435898, \"f1-score\": 0.6666666666666667, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7989348370927318, \"recall\": 0.8121539918414918, \"f1-score\": 0.7989948539871141, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7548801296169717, \"recall\": 0.7407407407407407, \"f1-score\": 0.7403936237579168, \"support\": 135.0}}", + "eval_f1_macro": 79.8994853987114, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.03936237579168, + "eval_loss": 1.1364786624908447, + "eval_runtime": 2.1228, + "eval_samples_per_second": 63.596, + "step": 2057 + }, + { + "epoch": 122.0, + "learning_rate": 0.00013, + "loss": 0.0011, + "step": 2074 + }, + { + "epoch": 122.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 74.26085371819231, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.875, \"f1-score\": 0.6363636363636364, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7922449945887446, \"recall\": 0.8090156371406372, \"f1-score\": 0.7898259455409633, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7544005531505531, \"recall\": 0.725925925925926, \"f1-score\": 0.728756351334877, \"support\": 135.0}}", + "eval_f1_macro": 78.98259455409632, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.87563513348772, + "eval_loss": 1.221017599105835, + "eval_runtime": 2.2505, + "eval_samples_per_second": 59.985, + "step": 2074 + }, + { + "epoch": 123.0, + "learning_rate": 0.00012833333333333333, + "loss": 0.0008, + "step": 2091 + }, + { + "epoch": 123.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.57727733650582, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.875, \"f1-score\": 0.6363636363636364, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7952244116037219, \"recall\": 0.7358318764568764, \"f1-score\": 0.7438368834916149, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7506233655658944, \"recall\": 0.7185185185185186, \"f1-score\": 0.7222171729315804, \"support\": 135.0}}", + "eval_f1_macro": 74.38368834916149, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.22171729315805, + "eval_loss": 1.2404636144638062, + "eval_runtime": 2.218, + "eval_samples_per_second": 60.867, + "step": 2091 + }, + { + "epoch": 124.0, + "learning_rate": 0.00012666666666666666, + "loss": 0.0048, + "step": 2108 + }, + { + "epoch": 124.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.91986381095504, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7971842356013552, \"recall\": 0.8078804875679876, \"f1-score\": 0.7945727056980254, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7516616475312128, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355551800735092, \"support\": 135.0}}", + "eval_f1_macro": 79.45727056980255, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.55551800735093, + "eval_loss": 1.1756157875061035, + "eval_runtime": 2.1659, + "eval_samples_per_second": 62.329, + "step": 2108 + }, + { + "epoch": 125.0, + "learning_rate": 0.000125, + "loss": 0.0007, + "step": 2125 + }, + { + "epoch": 125.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.36438346817907, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7759566186985541, \"recall\": 0.7887043512043512, \"f1-score\": 0.7763513430179385, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7347037992199282, \"recall\": 0.7185185185185186, \"f1-score\": 0.721186958672187, \"support\": 135.0}}", + "eval_f1_macro": 77.63513430179385, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.11869586721869, + "eval_loss": 1.212998628616333, + "eval_runtime": 2.2937, + "eval_samples_per_second": 58.857, + "step": 2125 + }, + { + "epoch": 126.0, + "learning_rate": 0.0001233333333333333, + "loss": 0.0008, + "step": 2142 + }, + { + "epoch": 126.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 71.27406229672498, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.5555555555555556, \"f1-score\": 0.6153846153846154, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7571729675177952, \"recall\": 0.7828950563325563, \"f1-score\": 0.7595188378961195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7231066799904603, \"recall\": 0.6962962962962963, \"f1-score\": 0.6988510613802864, \"support\": 135.0}}", + "eval_f1_macro": 75.95188378961197, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.88510613802865, + "eval_loss": 1.2854887247085571, + "eval_runtime": 2.074, + "eval_samples_per_second": 65.093, + "step": 2142 + }, + { + "epoch": 127.0, + "learning_rate": 0.00012166666666666665, + "loss": 0.0004, + "step": 2159 + }, + { + "epoch": 127.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.4292424329562, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7681011118511119, \"recall\": 0.7236001845376845, \"f1-score\": 0.72428547901374, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7304041637374971, \"recall\": 0.6962962962962963, \"f1-score\": 0.7002916257119156, \"support\": 135.0}}", + "eval_f1_macro": 72.42854790137399, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 70.02916257119156, + "eval_loss": 1.351884365081787, + "eval_runtime": 2.1241, + "eval_samples_per_second": 63.557, + "step": 2159 + }, + { + "epoch": 128.0, + "learning_rate": 0.00011999999999999999, + "loss": 0.0002, + "step": 2176 + }, + { + "epoch": 128.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.41850794610482, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7678106428106428, \"recall\": 0.7236001845376845, \"f1-score\": 0.7248628871217477, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7268032082846898, \"recall\": 0.6962962962962963, \"f1-score\": 0.6992848381298528, \"support\": 135.0}}", + "eval_f1_macro": 72.48628871217477, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.92848381298528, + "eval_loss": 1.3572306632995605, + "eval_runtime": 2.1876, + "eval_samples_per_second": 61.71, + "step": 2176 + }, + { + "epoch": 129.0, + "learning_rate": 0.00011833333333333331, + "loss": 0.001, + "step": 2193 + }, + { + "epoch": 129.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.61782293955142, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8034016636957814, \"recall\": 0.752324932012432, \"f1-score\": 0.7609662631209947, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564920917862095, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422651729795805, \"support\": 135.0}}", + "eval_f1_macro": 76.09662631209946, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.22651729795805, + "eval_loss": 1.2295937538146973, + "eval_runtime": 2.2337, + "eval_samples_per_second": 60.438, + "step": 2193 + }, + { + "epoch": 130.0, + "learning_rate": 0.00011666666666666665, + "loss": 0.0028, + "step": 2210 + }, + { + "epoch": 130.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.55886340237367, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7957131716322893, \"recall\": 0.8180300602175602, \"f1-score\": 0.7989535791265285, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7572416508691019, \"recall\": 0.7407407407407407, \"f1-score\": 0.7419194754869369, \"support\": 135.0}}", + "eval_f1_macro": 79.89535791265286, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.1919475486937, + "eval_loss": 1.259170651435852, + "eval_runtime": 2.1066, + "eval_samples_per_second": 64.083, + "step": 2210 + }, + { + "epoch": 131.0, + "learning_rate": 0.000115, + "loss": 0.0033, + "step": 2227 + }, + { + "epoch": 131.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 72.25202775378126, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7837695868945869, \"recall\": 0.7942586926961928, \"f1-score\": 0.7756004567204229, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7381803313284795, \"recall\": 0.7037037037037037, \"f1-score\": 0.7070732460234199, \"support\": 135.0}}", + "eval_f1_macro": 77.56004567204229, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.707324602342, + "eval_loss": 1.3266923427581787, + "eval_runtime": 2.1701, + "eval_samples_per_second": 62.208, + "step": 2227 + }, + { + "epoch": 132.0, + "learning_rate": 0.00011333333333333331, + "loss": 0.0004, + "step": 2244 + }, + { + "epoch": 132.0, + "eval_accuracy": 69.62962962962963, + "eval_average_metrics": 70.6257688194083, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7814234907984907, \"recall\": 0.7317586926961928, \"f1-score\": 0.7328372988256859, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327582425730573, \"recall\": 0.6962962962962963, \"f1-score\": 0.699600861358053, \"support\": 135.0}}", + "eval_f1_macro": 73.28372988256861, + "eval_f1_micro": 69.62962962962963, + "eval_f1_weighted": 69.9600861358053, + "eval_loss": 1.33100426197052, + "eval_runtime": 2.1896, + "eval_samples_per_second": 61.654, + "step": 2244 + }, + { + "epoch": 133.0, + "learning_rate": 0.00011166666666666667, + "loss": 0.0006, + "step": 2261 + }, + { + "epoch": 133.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.70429974221537, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7960415628762403, \"recall\": 0.7419082653457654, \"f1-score\": 0.7506471755169501, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7390933483406602, \"recall\": 0.7185185185185186, \"f1-score\": 0.7204877771346277, \"support\": 135.0}}", + "eval_f1_macro": 75.06471755169501, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.04877771346277, + "eval_loss": 1.2336255311965942, + "eval_runtime": 2.2249, + "eval_samples_per_second": 60.678, + "step": 2261 + }, + { + "epoch": 134.0, + "learning_rate": 0.00010999999999999998, + "loss": 0.0001, + "step": 2278 + }, + { + "epoch": 134.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 72.02848654875856, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7946632071632072, \"recall\": 0.7381689491064491, \"f1-score\": 0.7445463256662919, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7448381796529944, \"recall\": 0.7111111111111111, \"f1-score\": 0.7143709140618288, \"support\": 135.0}}", + "eval_f1_macro": 74.45463256662919, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.43709140618289, + "eval_loss": 1.3270906209945679, + "eval_runtime": 2.1891, + "eval_samples_per_second": 61.669, + "step": 2278 + }, + { + "epoch": 135.0, + "learning_rate": 0.00010833333333333333, + "loss": 0.0014, + "step": 2295 + }, + { + "epoch": 135.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.72547585668991, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6944444444444444, \"f1-score\": 0.7042253521126761, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8124975198412698, \"recall\": 0.7569323038073038, \"f1-score\": 0.7644743043935498, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7664916225749558, \"recall\": 0.7407407407407407, \"f1-score\": 0.7430632483925649, \"support\": 135.0}}", + "eval_f1_macro": 76.44743043935497, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.30632483925649, + "eval_loss": 1.251107931137085, + "eval_runtime": 2.1766, + "eval_samples_per_second": 62.023, + "step": 2295 + }, + { + "epoch": 136.0, + "learning_rate": 0.00010666666666666667, + "loss": 0.0024, + "step": 2312 + }, + { + "epoch": 136.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 76.36714659524762, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6842105263157895, \"recall\": 0.7222222222222222, \"f1-score\": 0.7027027027027027, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.8174468138737649, \"recall\": 0.8196993978243978, \"f1-score\": 0.8082919113289957, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7746230198740786, \"recall\": 0.7481481481481481, \"f1-score\": 0.7500976561846127, \"support\": 135.0}}", + "eval_f1_macro": 80.82919113289957, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 75.00976561846126, + "eval_loss": 1.2201787233352661, + "eval_runtime": 2.1085, + "eval_samples_per_second": 64.026, + "step": 2312 + }, + { + "epoch": 137.0, + "learning_rate": 0.00010499999999999999, + "loss": 0.0006, + "step": 2329 + }, + { + "epoch": 137.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 73.34865168139474, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.5277777777777778, \"f1-score\": 0.6229508196721312, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7731202865761689, \"recall\": 0.8114194832944832, \"f1-score\": 0.7785896412055908, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7473708731355789, \"recall\": 0.7185185185185186, \"f1-score\": 0.7183193890131623, \"support\": 135.0}}", + "eval_f1_macro": 77.85896412055907, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 71.83193890131622, + "eval_loss": 1.2545726299285889, + "eval_runtime": 2.0734, + "eval_samples_per_second": 65.11, + "step": 2329 + }, + { + "epoch": 138.0, + "learning_rate": 0.00010333333333333333, + "loss": 0.0018, + "step": 2346 + }, + { + "epoch": 138.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.43196457326891, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7872411405020101, \"recall\": 0.809604458041958, \"f1-score\": 0.792512077294686, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7577424186119838, \"recall\": 0.7407407407407407, \"f1-score\": 0.7432850241545893, \"support\": 135.0}}", + "eval_f1_macro": 79.2512077294686, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.32850241545893, + "eval_loss": 1.1618794202804565, + "eval_runtime": 2.1519, + "eval_samples_per_second": 62.735, + "step": 2346 + }, + { + "epoch": 139.0, + "learning_rate": 0.00010166666666666667, + "loss": 0.0004, + "step": 2363 + }, + { + "epoch": 139.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.73814060272393, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7714285714285715, \"recall\": 0.6923076923076923, \"f1-score\": 0.7297297297297296, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7857676140284836, \"recall\": 0.7979737276612278, \"f1-score\": 0.7869864182364182, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7477583286278938, \"recall\": 0.7333333333333333, \"f1-score\": 0.7358725392058726, \"support\": 135.0}}", + "eval_f1_macro": 78.69864182364182, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.58725392058726, + "eval_loss": 1.1620222330093384, + "eval_runtime": 2.2656, + "eval_samples_per_second": 59.586, + "step": 2363 + }, + { + "epoch": 140.0, + "learning_rate": 9.999999999999999e-05, + "loss": 0.0044, + "step": 2380 + }, + { + "epoch": 140.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.40867003367005, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.5833333333333334, \"f1-score\": 0.6363636363636365, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7788792530568847, \"recall\": 0.7317344114219114, \"f1-score\": 0.7391414141414141, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7357835666607596, \"recall\": 0.7185185185185186, \"f1-score\": 0.7201683501683502, \"support\": 135.0}}", + "eval_f1_macro": 73.91414141414143, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.01683501683502, + "eval_loss": 1.2089694738388062, + "eval_runtime": 2.2372, + "eval_samples_per_second": 60.342, + "step": 2380 + }, + { + "epoch": 141.0, + "learning_rate": 9.833333333333333e-05, + "loss": 0.0005, + "step": 2397 + }, + { + "epoch": 141.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.0998434515311, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.780597947085016, \"recall\": 0.7430980477855478, \"f1-score\": 0.7443116919947583, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7462899681290487, \"recall\": 0.725925925925926, \"f1-score\": 0.7278301942146337, \"support\": 135.0}}", + "eval_f1_macro": 74.43116919947583, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.78301942146337, + "eval_loss": 1.2010650634765625, + "eval_runtime": 2.2981, + "eval_samples_per_second": 58.744, + "step": 2397 + }, + { + "epoch": 142.0, + "learning_rate": 9.666666666666667e-05, + "loss": 0.0001, + "step": 2414 + }, + { + "epoch": 142.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 74.19280926182283, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7850638100067613, \"recall\": 0.8105514277389277, \"f1-score\": 0.7877778983036839, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7492036661407858, \"recall\": 0.725925925925926, \"f1-score\": 0.7280826203173772, \"support\": 135.0}}", + "eval_f1_macro": 78.77778983036839, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.80826203173773, + "eval_loss": 1.2088888883590698, + "eval_runtime": 2.1385, + "eval_samples_per_second": 63.128, + "step": 2414 + }, + { + "epoch": 143.0, + "learning_rate": 9.499999999999999e-05, + "loss": 0.001, + "step": 2431 + }, + { + "epoch": 143.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.53753615871489, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.793443807141625, \"recall\": 0.8174958721833722, \"f1-score\": 0.797105878846315, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7597858193999876, \"recall\": 0.7407407407407407, \"f1-score\": 0.7429140860207989, \"support\": 135.0}}", + "eval_f1_macro": 79.7105878846315, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.29140860207988, + "eval_loss": 1.1494632959365845, + "eval_runtime": 2.0877, + "eval_samples_per_second": 64.663, + "step": 2431 + }, + { + "epoch": 144.0, + "learning_rate": 9.333333333333333e-05, + "loss": 0.0003, + "step": 2448 + }, + { + "epoch": 144.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.5783020422968, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6944444444444444, \"f1-score\": 0.7042253521126761, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7968614718614719, \"recall\": 0.8134226884226884, \"f1-score\": 0.7990512962136822, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7549655282988617, \"recall\": 0.7407407407407407, \"f1-score\": 0.742599303996708, \"support\": 135.0}}", + "eval_f1_macro": 79.90512962136823, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.25993039967081, + "eval_loss": 1.1091960668563843, + "eval_runtime": 2.1685, + "eval_samples_per_second": 62.254, + "step": 2448 + }, + { + "epoch": 145.0, + "learning_rate": 9.166666666666667e-05, + "loss": 0.0002, + "step": 2465 + }, + { + "epoch": 145.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 76.29216423947909, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.8073717948717949, \"recall\": 0.8171620046620047, \"f1-score\": 0.8057793565410674, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7656030389363723, \"recall\": 0.7481481481481481, \"f1-score\": 0.7496109167418001, \"support\": 135.0}}", + "eval_f1_macro": 80.57793565410674, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 74.96109167418001, + "eval_loss": 1.1161596775054932, + "eval_runtime": 2.1314, + "eval_samples_per_second": 63.338, + "step": 2465 + }, + { + "epoch": 146.0, + "learning_rate": 8.999999999999999e-05, + "loss": 0.0001, + "step": 2482 + }, + { + "epoch": 146.0, + "eval_accuracy": 75.55555555555556, + "eval_average_metrics": 76.97719093283912, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8120300751879699, \"recall\": 0.8249745046620047, \"f1-score\": 0.8115000772617881, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7727188341223429, \"recall\": 0.7555555555555555, \"f1-score\": 0.7564764489406657, \"support\": 135.0}}", + "eval_f1_macro": 81.15000772617881, + "eval_f1_micro": 75.55555555555556, + "eval_f1_weighted": 75.64764489406657, + "eval_loss": 1.1192935705184937, + "eval_runtime": 2.1466, + "eval_samples_per_second": 62.891, + "step": 2482 + }, + { + "epoch": 147.0, + "learning_rate": 8.833333333333333e-05, + "loss": 0.0004, + "step": 2499 + }, + { + "epoch": 147.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.58963733628248, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7987710818593171, \"recall\": 0.809604458041958, \"f1-score\": 0.7987620895510332, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7557962734433322, \"recall\": 0.7407407407407407, \"f1-score\": 0.7433419224187846, \"support\": 135.0}}", + "eval_f1_macro": 79.87620895510332, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.33419224187845, + "eval_loss": 1.14371657371521, + "eval_runtime": 2.2241, + "eval_samples_per_second": 60.698, + "step": 2499 + }, + { + "epoch": 148.0, + "learning_rate": 8.666666666666665e-05, + "loss": 0.0006, + "step": 2516 + }, + { + "epoch": 148.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.05647965158836, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7792046221322537, \"recall\": 0.7430980477855478, \"f1-score\": 0.7418917435221783, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7518493527265457, \"recall\": 0.725925925925926, \"f1-score\": 0.7285155906895038, \"support\": 135.0}}", + "eval_f1_macro": 74.18917435221783, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.85155906895038, + "eval_loss": 1.2226479053497314, + "eval_runtime": 2.2006, + "eval_samples_per_second": 61.347, + "step": 2516 + }, + { + "epoch": 149.0, + "learning_rate": 8.499999999999999e-05, + "loss": 0.0001, + "step": 2533 + }, + { + "epoch": 149.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.12022510871195, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7809471629514733, \"recall\": 0.7433651418026419, \"f1-score\": 0.7449589518668466, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7465120614545901, \"recall\": 0.725925925925926, \"f1-score\": 0.7279982006297796, \"support\": 135.0}}", + "eval_f1_macro": 74.49589518668466, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.79982006297794, + "eval_loss": 1.2157655954360962, + "eval_runtime": 2.1514, + "eval_samples_per_second": 62.749, + "step": 2533 + }, + { + "epoch": 150.0, + "learning_rate": 8.333333333333333e-05, + "loss": 0.0009, + "step": 2550 + }, + { + "epoch": 150.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 76.0197940520429, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7942298714357537, \"recall\": 0.8026600135975136, \"f1-score\": 0.7949344892838044, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7581013031993423, \"recall\": 0.7481481481481481, \"f1-score\": 0.7495609765016158, \"support\": 135.0}}", + "eval_f1_macro": 79.49344892838043, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 74.95609765016158, + "eval_loss": 1.1733694076538086, + "eval_runtime": 2.2975, + "eval_samples_per_second": 58.759, + "step": 2550 + }, + { + "epoch": 151.0, + "learning_rate": 8.166666666666665e-05, + "loss": 0.0008, + "step": 2567 + }, + { + "epoch": 151.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.8567648637555, + "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7776136152494848, \"recall\": 0.7818266802641802, \"f1-score\": 0.7749303779482808, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7380173905536225, \"recall\": 0.725925925925926, \"f1-score\": 0.7274883647500868, \"support\": 135.0}}", + "eval_f1_macro": 77.49303779482808, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.74883647500869, + "eval_loss": 1.1846721172332764, + "eval_runtime": 2.2205, + "eval_samples_per_second": 60.796, + "step": 2567 + }, + { + "epoch": 152.0, + "learning_rate": 7.999999999999999e-05, + "loss": 0.0001, + "step": 2584 + }, + { + "epoch": 152.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 75.57891605906845, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8020374552983248, \"recall\": 0.8070792055167055, \"f1-score\": 0.7986914184004976, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.755954576824142, \"recall\": 0.7407407407407407, \"f1-score\": 0.7429837424807587, \"support\": 135.0}}", + "eval_f1_macro": 79.86914184004976, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.29837424807587, + "eval_loss": 1.188364863395691, + "eval_runtime": 2.1962, + "eval_samples_per_second": 61.471, + "step": 2584 + }, + { + "epoch": 153.0, + "learning_rate": 7.833333333333333e-05, + "loss": 0.0001, + "step": 2601 + }, + { + "epoch": 153.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.21584842556148, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6744186046511628, \"recall\": 0.7435897435897436, \"f1-score\": 0.7073170731707318, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8045674770965469, \"recall\": 0.7175905691530691, \"f1-score\": 0.7470421497111972, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7428455093571373, \"recall\": 0.725925925925926, \"f1-score\": 0.7297399354594096, \"support\": 135.0}}", + "eval_f1_macro": 74.70421497111973, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.97399354594096, + "eval_loss": 1.2244882583618164, + "eval_runtime": 2.2128, + "eval_samples_per_second": 61.008, + "step": 2601 + }, + { + "epoch": 154.0, + "learning_rate": 7.666666666666666e-05, + "loss": 0.0039, + "step": 2618 + }, + { + "epoch": 154.0, + "eval_accuracy": 75.55555555555556, + "eval_average_metrics": 76.95897870855806, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7647058823529411, \"recall\": 0.7222222222222222, \"f1-score\": 0.7428571428571428, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8151371074745688, \"recall\": 0.8136897824397824, \"f1-score\": 0.8096817403891035, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.766684030337281, \"recall\": 0.7555555555555555, \"f1-score\": 0.7575662968421079, \"support\": 135.0}}", + "eval_f1_macro": 80.96817403891035, + "eval_f1_micro": 75.55555555555556, + "eval_f1_weighted": 75.75662968421078, + "eval_loss": 1.2217811346054077, + "eval_runtime": 2.2436, + "eval_samples_per_second": 60.172, + "step": 2618 + }, + { + "epoch": 155.0, + "learning_rate": 7.5e-05, + "loss": 0.0003, + "step": 2635 + }, + { + "epoch": 155.0, + "eval_accuracy": 75.55555555555556, + "eval_average_metrics": 77.04653690264276, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8132034632034633, \"recall\": 0.8244403166278166, \"f1-score\": 0.8132795190869966, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7693025493025493, \"recall\": 0.7555555555555555, \"f1-score\": 0.7574708459076027, \"support\": 135.0}}", + "eval_f1_macro": 81.32795190869966, + "eval_f1_micro": 75.55555555555556, + "eval_f1_weighted": 75.74708459076027, + "eval_loss": 1.2007545232772827, + "eval_runtime": 2.1766, + "eval_samples_per_second": 62.023, + "step": 2635 + }, + { + "epoch": 156.0, + "learning_rate": 7.333333333333332e-05, + "loss": 0.0002, + "step": 2652 + }, + { + "epoch": 156.0, + "eval_accuracy": 75.55555555555556, + "eval_average_metrics": 77.04477692135627, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8131493506493507, \"recall\": 0.8244403166278166, \"f1-score\": 0.8131887812803564, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7696726367314601, \"recall\": 0.7555555555555555, \"f1-score\": 0.7574911844627835, \"support\": 135.0}}", + "eval_f1_macro": 81.31887812803565, + "eval_f1_micro": 75.55555555555556, + "eval_f1_weighted": 75.74911844627835, + "eval_loss": 1.1942009925842285, + "eval_runtime": 2.3622, + "eval_samples_per_second": 57.15, + "step": 2652 + }, + { + "epoch": 157.0, + "learning_rate": 7.166666666666667e-05, + "loss": 0.0022, + "step": 2669 + }, + { + "epoch": 157.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.09230915854165, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8025137608806963, \"recall\": 0.7549958721833722, \"f1-score\": 0.7614065508371455, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7533344711839336, \"recall\": 0.7333333333333333, \"f1-score\": 0.735619148837854, \"support\": 135.0}}", + "eval_f1_macro": 76.14065508371455, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.56191488378542, + "eval_loss": 1.217727541923523, + "eval_runtime": 2.2066, + "eval_samples_per_second": 61.18, + "step": 2669 + }, + { + "epoch": 158.0, + "learning_rate": 7e-05, + "loss": 0.0, + "step": 2686 + }, + { + "epoch": 158.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.62519417356738, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", + "eval_f1_macro": 74.72856934535828, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06850364520756, + "eval_loss": 1.2388432025909424, + "eval_runtime": 2.153, + "eval_samples_per_second": 62.703, + "step": 2686 + }, + { + "epoch": 159.0, + "learning_rate": 6.833333333333333e-05, + "loss": 0.0006, + "step": 2703 + }, + { + "epoch": 159.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.02792297951471, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040257260365956, \"recall\": 0.7488527097902098, \"f1-score\": 0.7588762507291247, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7526372335067987, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355740017847974, \"support\": 135.0}}", + "eval_f1_macro": 75.88762507291247, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.55740017847972, + "eval_loss": 1.221247673034668, + "eval_runtime": 2.273, + "eval_samples_per_second": 59.392, + "step": 2703 + }, + { + "epoch": 160.0, + "learning_rate": 6.666666666666666e-05, + "loss": 0.0028, + "step": 2720 + }, + { + "epoch": 160.0, + "eval_accuracy": 74.81481481481481, + "eval_average_metrics": 76.18622883029525, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7428571428571429, \"recall\": 0.7222222222222222, \"f1-score\": 0.732394366197183, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7960499178241114, \"recall\": 0.8215022824397824, \"f1-score\": 0.8033114500290621, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7596155099380906, \"recall\": 0.7481481481481481, \"f1-score\": 0.7478414068864512, \"support\": 135.0}}", + "eval_f1_macro": 80.33114500290621, + "eval_f1_micro": 74.81481481481481, + "eval_f1_weighted": 74.78414068864512, + "eval_loss": 1.1919903755187988, + "eval_runtime": 2.2082, + "eval_samples_per_second": 61.136, + "step": 2720 + }, + { + "epoch": 161.0, + "learning_rate": 6.5e-05, + "loss": 0.0005, + "step": 2737 + }, + { + "epoch": 161.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.24915577342048, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7860111205219901, \"recall\": 0.75152364996115, \"f1-score\": 0.7518137254901962, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7415646457675443, \"recall\": 0.725925925925926, \"f1-score\": 0.7263006535947714, \"support\": 135.0}}", + "eval_f1_macro": 75.1813725490196, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.63006535947713, + "eval_loss": 1.2137112617492676, + "eval_runtime": 2.2438, + "eval_samples_per_second": 60.166, + "step": 2737 + }, + { + "epoch": 162.0, + "learning_rate": 6.333333333333333e-05, + "loss": 0.0003, + "step": 2754 + }, + { + "epoch": 162.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 74.01724012020507, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.797503056005861, \"recall\": 0.7549958721833722, \"f1-score\": 0.7588846210125841, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7518832439356048, \"recall\": 0.7333333333333333, \"f1-score\": 0.7351383171289522, \"support\": 135.0}}", + "eval_f1_macro": 75.88846210125841, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.51383171289521, + "eval_loss": 1.2145878076553345, + "eval_runtime": 2.142, + "eval_samples_per_second": 63.026, + "step": 2754 + }, + { + "epoch": 163.0, + "learning_rate": 6.166666666666666e-05, + "loss": 0.0007, + "step": 2771 + }, + { + "epoch": 163.0, + "eval_accuracy": 74.07407407407408, + "eval_average_metrics": 74.5749463235295, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7027027027027027, \"recall\": 0.7222222222222222, \"f1-score\": 0.7123287671232876, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.797979519073269, \"recall\": 0.7508437742812742, \"f1-score\": 0.7594312360087268, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.755945511362178, \"recall\": 0.7407407407407407, \"f1-score\": 0.7420851354509721, \"support\": 135.0}}", + "eval_f1_macro": 75.94312360087267, + "eval_f1_micro": 74.07407407407408, + "eval_f1_weighted": 74.2085135450972, + "eval_loss": 1.2377903461456299, + "eval_runtime": 2.1826, + "eval_samples_per_second": 61.852, + "step": 2771 + }, + { + "epoch": 164.0, + "learning_rate": 5.9999999999999995e-05, + "loss": 0.0012, + "step": 2788 + }, + { + "epoch": 164.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.9473371413507, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7955596194087242, \"recall\": 0.747104458041958, \"f1-score\": 0.7549024404282261, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7527513312679553, \"recall\": 0.7333333333333333, \"f1-score\": 0.7363243785591355, \"support\": 135.0}}", + "eval_f1_macro": 75.49024404282261, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.63243785591355, + "eval_loss": 1.2320207357406616, + "eval_runtime": 2.1945, + "eval_samples_per_second": 61.519, + "step": 2788 + }, + { + "epoch": 165.0, + "learning_rate": 5.8333333333333326e-05, + "loss": 0.0002, + "step": 2805 + }, + { + "epoch": 165.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.35054466230937, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7934839466089466, \"recall\": 0.75152364996115, \"f1-score\": 0.7543137254901962, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7467981401314735, \"recall\": 0.725925925925926, \"f1-score\": 0.7278562091503267, \"support\": 135.0}}", + "eval_f1_macro": 75.43137254901961, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.78562091503268, + "eval_loss": 1.2451300621032715, + "eval_runtime": 2.1694, + "eval_samples_per_second": 62.23, + "step": 2805 + }, + { + "epoch": 166.0, + "learning_rate": 5.666666666666666e-05, + "loss": 0.0004, + "step": 2822 + }, + { + "epoch": 166.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.29745663811057, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7926096513780337, \"recall\": 0.7479725135975136, \"f1-score\": 0.7521030279803201, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7472310129957188, \"recall\": 0.725925925925926, \"f1-score\": 0.7279433856922508, \"support\": 135.0}}", + "eval_f1_macro": 75.21030279803202, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.79433856922508, + "eval_loss": 1.2745343446731567, + "eval_runtime": 2.2435, + "eval_samples_per_second": 60.174, + "step": 2822 + }, + { + "epoch": 167.0, + "learning_rate": 5.499999999999999e-05, + "loss": 0.0001, + "step": 2839 + }, + { + "epoch": 167.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.96219025098748, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7966899766899767, \"recall\": 0.7514447358197358, \"f1-score\": 0.7557906902598437, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564506604506605, \"recall\": 0.7333333333333333, \"f1-score\": 0.7360302531129886, \"support\": 135.0}}", + "eval_f1_macro": 75.57906902598438, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.60302531129886, + "eval_loss": 1.2781518697738647, + "eval_runtime": 2.2619, + "eval_samples_per_second": 59.684, + "step": 2839 + }, + { + "epoch": 168.0, + "learning_rate": 5.333333333333333e-05, + "loss": 0.0002, + "step": 2856 + }, + { + "epoch": 168.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.56451583121837, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7900713712810488, \"recall\": 0.738357128982129, \"f1-score\": 0.7450481168947851, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740871934119246, \"recall\": 0.7185185185185186, \"f1-score\": 0.7204954793169126, \"support\": 135.0}}", + "eval_f1_macro": 74.50481168947852, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.04954793169127, + "eval_loss": 1.305252194404602, + "eval_runtime": 2.1319, + "eval_samples_per_second": 63.325, + "step": 2856 + }, + { + "epoch": 169.0, + "learning_rate": 5.1666666666666664e-05, + "loss": 0.0001, + "step": 2873 + }, + { + "epoch": 169.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89929243227385, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", + "eval_f1_macro": 73.97919123590555, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.39575627096765, + "eval_loss": 1.3155133724212646, + "eval_runtime": 2.2317, + "eval_samples_per_second": 60.491, + "step": 2873 + }, + { + "epoch": 170.0, + "learning_rate": 4.9999999999999996e-05, + "loss": 0.0, + "step": 2890 + }, + { + "epoch": 170.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89929243227385, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", + "eval_f1_macro": 73.97919123590555, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.39575627096765, + "eval_loss": 1.3125479221343994, + "eval_runtime": 2.1736, + "eval_samples_per_second": 62.109, + "step": 2890 + }, + { + "epoch": 171.0, + "learning_rate": 4.8333333333333334e-05, + "loss": 0.0001, + "step": 2907 + }, + { + "epoch": 171.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89929243227385, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", + "eval_f1_macro": 73.97919123590555, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.39575627096765, + "eval_loss": 1.3118623495101929, + "eval_runtime": 2.1217, + "eval_samples_per_second": 63.628, + "step": 2907 + }, + { + "epoch": 172.0, + "learning_rate": 4.6666666666666665e-05, + "loss": 0.0, + "step": 2924 + }, + { + "epoch": 172.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89929243227385, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", + "eval_f1_macro": 73.97919123590555, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.39575627096765, + "eval_loss": 1.312318205833435, + "eval_runtime": 2.0073, + "eval_samples_per_second": 67.255, + "step": 2924 + }, + { + "epoch": 173.0, + "learning_rate": 4.4999999999999996e-05, + "loss": 0.0, + "step": 2941 + }, + { + "epoch": 173.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89929243227385, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", + "eval_f1_macro": 73.97919123590555, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.39575627096765, + "eval_loss": 1.3142341375350952, + "eval_runtime": 2.155, + "eval_samples_per_second": 62.645, + "step": 2941 + }, + { + "epoch": 174.0, + "learning_rate": 4.333333333333333e-05, + "loss": 0.0001, + "step": 2958 + }, + { + "epoch": 174.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.9747073753716, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040708376645876, \"recall\": 0.7455686674436675, \"f1-score\": 0.7566759148589195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7558651279484612, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356457134892781, \"support\": 135.0}}", + "eval_f1_macro": 75.66759148589195, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.5645713489278, + "eval_loss": 1.2926682233810425, + "eval_runtime": 2.0992, + "eval_samples_per_second": 64.311, + "step": 2958 + }, + { + "epoch": 175.0, + "learning_rate": 4.1666666666666665e-05, + "loss": 0.0001, + "step": 2975 + }, + { + "epoch": 175.0, + "eval_accuracy": 73.33333333333333, + "eval_average_metrics": 73.9747073753716, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040708376645876, \"recall\": 0.7455686674436675, \"f1-score\": 0.7566759148589195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7558651279484612, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356457134892781, \"support\": 135.0}}", + "eval_f1_macro": 75.66759148589195, + "eval_f1_micro": 73.33333333333333, + "eval_f1_weighted": 73.5645713489278, + "eval_loss": 1.2941668033599854, + "eval_runtime": 2.1293, + "eval_samples_per_second": 63.402, + "step": 2975 + }, + { + "epoch": 176.0, + "learning_rate": 3.9999999999999996e-05, + "loss": 0.0002, + "step": 2992 + }, + { + "epoch": 176.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.61480524125062, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7935292547718269, \"recall\": 0.7337497571872572, \"f1-score\": 0.7460040211884412, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7400211319000227, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215511514245466, \"support\": 135.0}}", + "eval_f1_macro": 74.60040211884412, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.15511514245466, + "eval_loss": 1.3083090782165527, + "eval_runtime": 2.2303, + "eval_samples_per_second": 60.529, + "step": 2992 + }, + { + "epoch": 177.0, + "learning_rate": 3.833333333333333e-05, + "loss": 0.0002, + "step": 3009 + }, + { + "epoch": 177.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.60305283815627, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7928594842657343, \"recall\": 0.7294094794094794, \"f1-score\": 0.7450467847762234, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7376262626262626, \"recall\": 0.7185185185185186, \"f1-score\": 0.7220382917129899, \"support\": 135.0}}", + "eval_f1_macro": 74.50467847762235, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.203829171299, + "eval_loss": 1.3025906085968018, + "eval_runtime": 2.2827, + "eval_samples_per_second": 59.141, + "step": 3009 + }, + { + "epoch": 178.0, + "learning_rate": 3.666666666666666e-05, + "loss": 0.0001, + "step": 3026 + }, + { + "epoch": 178.0, + "eval_accuracy": 72.5925925925926, + "eval_average_metrics": 73.32833124918041, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.717948717948718, \"f1-score\": 0.708860759493671, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.80234375, \"recall\": 0.7326146076146076, \"f1-score\": 0.7515625584906763, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7454040404040404, \"recall\": 0.725925925925926, \"f1-score\": 0.7297188396246881, \"support\": 135.0}}", + "eval_f1_macro": 75.15625584906763, + "eval_f1_micro": 72.59259259259261, + "eval_f1_weighted": 72.97188396246881, + "eval_loss": 1.2811334133148193, + "eval_runtime": 2.3185, + "eval_samples_per_second": 58.228, + "step": 3026 + }, + { + "epoch": 179.0, + "learning_rate": 3.5e-05, + "loss": 0.002, + "step": 3043 + }, + { + "epoch": 179.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.94066572336109, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7887609848076188, \"recall\": 0.7259372571872572, \"f1-score\": 0.740553221987642, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7323961919333588, \"recall\": 0.7111111111111111, \"f1-score\": 0.7148511847245799, \"support\": 135.0}}", + "eval_f1_macro": 74.05532219876419, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.48511847245798, + "eval_loss": 1.3122072219848633, + "eval_runtime": 2.2232, + "eval_samples_per_second": 60.724, + "step": 3043 + }, + { + "epoch": 180.0, + "learning_rate": 3.333333333333333e-05, + "loss": 0.0001, + "step": 3060 + }, + { + "epoch": 180.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.62519417356738, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", + "eval_f1_macro": 74.72856934535828, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06850364520756, + "eval_loss": 1.3237799406051636, + "eval_runtime": 2.2054, + "eval_samples_per_second": 61.213, + "step": 3060 + }, + { + "epoch": 181.0, + "learning_rate": 3.1666666666666666e-05, + "loss": 0.0001, + "step": 3077 + }, + { + "epoch": 181.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.62519417356738, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", + "eval_f1_macro": 74.72856934535828, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06850364520756, + "eval_loss": 1.3265109062194824, + "eval_runtime": 2.3696, + "eval_samples_per_second": 56.971, + "step": 3077 + }, + { + "epoch": 182.0, + "learning_rate": 2.9999999999999997e-05, + "loss": 0.0, + "step": 3094 + }, + { + "epoch": 182.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.62519417356738, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", + "eval_f1_macro": 74.72856934535828, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06850364520756, + "eval_loss": 1.3261035680770874, + "eval_runtime": 2.2698, + "eval_samples_per_second": 59.477, + "step": 3094 + }, + { + "epoch": 183.0, + "learning_rate": 2.833333333333333e-05, + "loss": 0.0009, + "step": 3111 + }, + { + "epoch": 183.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.22624100757768, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7805373005970924, \"recall\": 0.722732128982129, \"f1-score\": 0.7343542579082432, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7253561559459545, \"recall\": 0.7037037037037037, \"f1-score\": 0.7072879749874571, \"support\": 135.0}}", + "eval_f1_macro": 73.43542579082431, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.7287974987457, + "eval_loss": 1.2959227561950684, + "eval_runtime": 2.1932, + "eval_samples_per_second": 61.554, + "step": 3111 + }, + { + "epoch": 184.0, + "learning_rate": 2.6666666666666667e-05, + "loss": 0.0, + "step": 3128 + }, + { + "epoch": 184.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.55073789502072, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7889478942110522, \"recall\": 0.7296765734265734, \"f1-score\": 0.7433792315665717, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7359373569899886, \"recall\": 0.7185185185185186, \"f1-score\": 0.7216132471972199, \"support\": 135.0}}", + "eval_f1_macro": 74.33792315665717, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.161324719722, + "eval_loss": 1.2773268222808838, + "eval_runtime": 2.157, + "eval_samples_per_second": 62.586, + "step": 3128 + }, + { + "epoch": 185.0, + "learning_rate": 2.4999999999999998e-05, + "loss": 0.0007, + "step": 3145 + }, + { + "epoch": 185.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.5453732812465, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7885856194679723, \"recall\": 0.7296765734265734, \"f1-score\": 0.7432477269156164, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7355754414577943, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215301672972063, \"support\": 135.0}}", + "eval_f1_macro": 74.32477269156163, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.15301672972065, + "eval_loss": 1.2745593786239624, + "eval_runtime": 2.2478, + "eval_samples_per_second": 60.059, + "step": 3145 + }, + { + "epoch": 186.0, + "learning_rate": 2.3333333333333332e-05, + "loss": 0.0001, + "step": 3162 + }, + { + "epoch": 186.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.5453732812465, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7885856194679723, \"recall\": 0.7296765734265734, \"f1-score\": 0.7432477269156164, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7355754414577943, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215301672972063, \"support\": 135.0}}", + "eval_f1_macro": 74.32477269156163, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.15301672972065, + "eval_loss": 1.273395299911499, + "eval_runtime": 2.1937, + "eval_samples_per_second": 61.539, + "step": 3162 + }, + { + "epoch": 187.0, + "learning_rate": 2.1666666666666664e-05, + "loss": 0.002, + "step": 3179 + }, + { + "epoch": 187.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89929243227385, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", + "eval_f1_macro": 73.97919123590555, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.39575627096765, + "eval_loss": 1.2829253673553467, + "eval_runtime": 2.2014, + "eval_samples_per_second": 61.325, + "step": 3179 + }, + { + "epoch": 188.0, + "learning_rate": 1.9999999999999998e-05, + "loss": 0.0001, + "step": 3196 + }, + { + "epoch": 188.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.62519417356738, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", + "eval_f1_macro": 74.72856934535828, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06850364520756, + "eval_loss": 1.3078842163085938, + "eval_runtime": 2.2833, + "eval_samples_per_second": 59.124, + "step": 3196 + }, + { + "epoch": 189.0, + "learning_rate": 1.833333333333333e-05, + "loss": 0.0, + "step": 3213 + }, + { + "epoch": 189.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.62216036599705, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}", + "eval_f1_macro": 74.71822632362485, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06671143665966, + "eval_loss": 1.2980728149414062, + "eval_runtime": 2.2356, + "eval_samples_per_second": 60.387, + "step": 3213 + }, + { + "epoch": 190.0, + "learning_rate": 1.6666666666666664e-05, + "loss": 0.0, + "step": 3230 + }, + { + "epoch": 190.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.62216036599705, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}", + "eval_f1_macro": 74.71822632362485, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06671143665966, + "eval_loss": 1.2942575216293335, + "eval_runtime": 2.1696, + "eval_samples_per_second": 62.223, + "step": 3230 + }, + { + "epoch": 191.0, + "learning_rate": 1.4999999999999999e-05, + "loss": 0.0001, + "step": 3247 + }, + { + "epoch": 191.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.974518342938, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7867627164502164, \"recall\": 0.7387031371406372, \"f1-score\": 0.7437678567314118, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7322831489498156, \"recall\": 0.7111111111111111, \"f1-score\": 0.7129906547638859, \"support\": 135.0}}", + "eval_f1_macro": 74.37678567314117, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.29906547638859, + "eval_loss": 1.293574571609497, + "eval_runtime": 2.1704, + "eval_samples_per_second": 62.2, + "step": 3247 + }, + { + "epoch": 192.0, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.0, + "step": 3264 + }, + { + "epoch": 192.0, + "eval_accuracy": 71.85185185185186, + "eval_average_metrics": 72.62216036599705, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}", + "eval_f1_macro": 74.71822632362485, + "eval_f1_micro": 71.85185185185186, + "eval_f1_weighted": 72.06671143665966, + "eval_loss": 1.2941248416900635, + "eval_runtime": 2.1999, + "eval_samples_per_second": 61.367, + "step": 3264 + }, + { + "epoch": 193.0, + "learning_rate": 1.1666666666666666e-05, + "loss": 0.0001, + "step": 3281 + }, + { + "epoch": 193.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89912419879386, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7850075271950272, \"recall\": 0.730544628982129, \"f1-score\": 0.739725063938619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327851161184494, \"recall\": 0.7111111111111111, \"f1-score\": 0.714017681790913, \"support\": 135.0}}", + "eval_f1_macro": 73.9725063938619, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.40176817909129, + "eval_loss": 1.2996220588684082, + "eval_runtime": 2.3185, + "eval_samples_per_second": 58.228, + "step": 3281 + }, + { + "epoch": 194.0, + "learning_rate": 9.999999999999999e-06, + "loss": 0.0001, + "step": 3298 + }, + { + "epoch": 194.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.89912419879386, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7850075271950272, \"recall\": 0.730544628982129, \"f1-score\": 0.739725063938619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327851161184494, \"recall\": 0.7111111111111111, \"f1-score\": 0.714017681790913, \"support\": 135.0}}", + "eval_f1_macro": 73.9725063938619, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.40176817909129, + "eval_loss": 1.3008208274841309, + "eval_runtime": 2.2664, + "eval_samples_per_second": 59.567, + "step": 3298 + }, + { + "epoch": 195.0, + "learning_rate": 8.333333333333332e-06, + "loss": 0.0001, + "step": 3315 + }, + { + "epoch": 195.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.2542231596473, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.781411297036297, \"recall\": 0.7273395007770008, \"f1-score\": 0.7363456487177715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7249367916034583, \"recall\": 0.7037037037037037, \"f1-score\": 0.7064158702607126, \"support\": 135.0}}", + "eval_f1_macro": 73.63456487177716, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.64158702607126, + "eval_loss": 1.2944679260253906, + "eval_runtime": 2.2091, + "eval_samples_per_second": 61.112, + "step": 3315 + }, + { + "epoch": 196.0, + "learning_rate": 6.666666666666667e-06, + "loss": 0.0154, + "step": 3332 + }, + { + "epoch": 196.0, + "eval_accuracy": 70.37037037037037, + "eval_average_metrics": 71.2542231596473, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.781411297036297, \"recall\": 0.7273395007770008, \"f1-score\": 0.7363456487177715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7249367916034583, \"recall\": 0.7037037037037037, \"f1-score\": 0.7064158702607126, \"support\": 135.0}}", + "eval_f1_macro": 73.63456487177716, + "eval_f1_micro": 70.37037037037037, + "eval_f1_weighted": 70.64158702607126, + "eval_loss": 1.2910572290420532, + "eval_runtime": 2.3432, + "eval_samples_per_second": 57.614, + "step": 3332 + }, + { + "epoch": 197.0, + "learning_rate": 4.9999999999999996e-06, + "loss": 0.0003, + "step": 3349 + }, + { + "epoch": 197.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.91662203626719, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}", + "eval_f1_macro": 74.08616942664258, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.35809649620393, + "eval_loss": 1.287298560142517, + "eval_runtime": 2.18, + "eval_samples_per_second": 61.926, + "step": 3349 + }, + { + "epoch": 198.0, + "learning_rate": 3.3333333333333333e-06, + "loss": 0.0001, + "step": 3366 + }, + { + "epoch": 198.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.91662203626719, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}", + "eval_f1_macro": 74.08616942664258, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.35809649620393, + "eval_loss": 1.2868810892105103, + "eval_runtime": 2.1185, + "eval_samples_per_second": 63.724, + "step": 3366 + }, + { + "epoch": 199.0, + "learning_rate": 1.6666666666666667e-06, + "loss": 0.0002, + "step": 3383 + }, + { + "epoch": 199.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.91662203626719, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}", + "eval_f1_macro": 74.08616942664258, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.35809649620393, + "eval_loss": 1.2866387367248535, + "eval_runtime": 2.1869, + "eval_samples_per_second": 61.732, + "step": 3383 + }, + { + "epoch": 200.0, + "learning_rate": 0.0, + "loss": 0.0081, + "step": 3400 + }, + { + "epoch": 200.0, + "eval_accuracy": 71.11111111111111, + "eval_average_metrics": 71.91662203626719, + "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}", + "eval_f1_macro": 74.08616942664258, + "eval_f1_micro": 71.11111111111111, + "eval_f1_weighted": 71.35809649620393, + "eval_loss": 1.2865550518035889, + "eval_runtime": 2.1903, + "eval_samples_per_second": 61.635, + "step": 3400 + }, + { + "epoch": 200.0, + "step": 3400, + "total_flos": 1.9410334099963904e+16, + "train_runtime": 2245.6861, + "train_samples_per_second": 1.514 + } + ], + "max_steps": 3400, + "num_train_epochs": 200, + "total_flos": 1.9410334099963904e+16, + "trial_name": null, + "trial_params": null +}