{ "best_metric": 77.22228406219811, "best_model_checkpoint": "/mounts/work/faeze/attempt/new_setting_outputs/source_adapter/crisis_8/100/checkpoint-1479", "epoch": 200.0, "global_step": 3400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 1.0, "learning_rate": 1.4999999999999999e-05, "loss": 9.5935, "step": 17 }, { "epoch": 1.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \")\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": D\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Hiding in\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification:\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 9.580382347106934, "eval_runtime": 2.2821, "eval_samples_per_second": 59.156, "step": 17 }, { "epoch": 2.0, "learning_rate": 2.9999999999999997e-05, "loss": 9.0928, "step": 34 }, { "epoch": 2.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"\\\"media\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \")\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"99.5%\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": apparently\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Emers\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Climate change is\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Hiding in\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No they have\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 9.190422058105469, "eval_runtime": 2.2408, "eval_samples_per_second": 60.245, "step": 34 }, { "epoch": 3.0, "learning_rate": 4.4999999999999996e-05, "loss": 8.1515, "step": 51 }, { "epoch": 3.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"-\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0.0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": Lots\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": please\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \": very\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"@Anders\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Can use passport\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"No Friday Im\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC 2019\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Shawn\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada is under\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification: climate\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"instagram\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 8.223907470703125, "eval_runtime": 2.3068, "eval_samples_per_second": 58.523, "step": 51 }, { "epoch": 4.0, "learning_rate": 5.9999999999999995e-05, "loss": 6.6637, "step": 68 }, { "epoch": 4.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\\\"How Dar\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0.0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Andrew Neil\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"False\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"True\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a friend\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a real\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"classification: good\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"panne\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"please show us\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"very goi\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 6.206799030303955, "eval_runtime": 2.267, "eval_samples_per_second": 59.549, "step": 68 }, { "epoch": 5.0, "learning_rate": 7.5e-05, "loss": 4.5728, "step": 85 }, { "epoch": 5.0, "eval_accuracy": 0.0, "eval_average_metrics": 0.0, "eval_classification_report": "{\"\\\"Reality\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \".\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"0\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"1\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"33000f\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \":\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"?\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"????\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Andrew Neil\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Australian\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Bruh\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"CNN\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"COVID\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Congress needs to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Ddx\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"Florida\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"India\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"PPC\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"SCP-04\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"a\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"canada\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 36.0}, \"dal\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"doubtful\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"eder\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"entail\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"if de\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"microwave\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"name\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"negative\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_d\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"not_en\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 16.0}, \"to\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"troll\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"trud\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"u\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"vac\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"weather\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.0, \"macro avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 135.0}}", "eval_f1_macro": 0.0, "eval_f1_micro": 0.0, "eval_f1_weighted": 0.0, "eval_loss": 3.01625657081604, "eval_runtime": 2.3046, "eval_samples_per_second": 58.579, "step": 85 }, { "epoch": 6.0, "learning_rate": 8.999999999999999e-05, "loss": 2.1563, "step": 102 }, { "epoch": 6.0, "eval_accuracy": 25.925925925925924, "eval_average_metrics": 18.478859129440288, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 11.0}, \"challenge\": {\"precision\": 0.27049180327868855, \"recall\": 0.9166666666666666, \"f1-score\": 0.4177215189873418, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 39.0}, \"not_e\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"positive\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.125, \"f1-score\": 0.2222222222222222, \"support\": 16.0}, \"statement\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"accuracy\": 0.25925925925925924, \"macro avg\": {\"precision\": 0.11549925484351714, \"recall\": 0.09469696969696968, \"f1-score\": 0.058176703746324, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.19064966605950215, \"recall\": 0.25925925925925924, \"f1-score\": 0.13772985362296192, \"support\": 135.0}}", "eval_f1_macro": 7.99929676511955, "eval_f1_micro": 26.217228464419474, "eval_f1_weighted": 13.772985362296192, "eval_loss": 1.1285314559936523, "eval_runtime": 2.054, "eval_samples_per_second": 65.725, "step": 102 }, { "epoch": 7.0, "learning_rate": 0.00010499999999999999, "loss": 1.0166, "step": 119 }, { "epoch": 7.0, "eval_accuracy": 40.0, "eval_average_metrics": 33.589515948427554, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"apology\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.2727272727272727, \"f1-score\": 0.39999999999999997, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.16666666666666666, \"f1-score\": 0.26666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.3392857142857143, \"recall\": 0.9743589743589743, \"f1-score\": 0.5033112582781457, \"support\": 39.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 13.0}, \"request\": {\"precision\": 1.0, \"recall\": 0.4375, \"f1-score\": 0.6086956521739131, \"support\": 16.0}, \"accuracy\": 0.4, \"macro avg\": {\"precision\": 0.3444940476190476, \"recall\": 0.23140661421911424, \"f1-score\": 0.22233419713984065, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.45542328042328045, \"recall\": 0.4, \"f1-score\": 0.3212464407972614, \"support\": 135.0}}", "eval_f1_macro": 22.233419713984066, "eval_f1_micro": 40.00000000000001, "eval_f1_weighted": 32.12464407972614, "eval_loss": 0.6725317239761353, "eval_runtime": 2.2366, "eval_samples_per_second": 60.359, "step": 119 }, { "epoch": 8.0, "learning_rate": 0.00011999999999999999, "loss": 0.7319, "step": 136 }, { "epoch": 8.0, "eval_accuracy": 48.888888888888886, "eval_average_metrics": 46.684741490249316, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.4, \"recall\": 0.2222222222222222, \"f1-score\": 0.2857142857142857, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.45454545454545453, \"f1-score\": 0.5555555555555556, \"support\": 11.0}, \"challenge\": {\"precision\": 0.4426229508196721, \"recall\": 0.75, \"f1-score\": 0.556701030927835, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.47619047619047616, \"recall\": 0.5128205128205128, \"f1-score\": 0.49382716049382713, \"support\": 39.0}, \"question\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 9.0}, \"rejection\": {\"precision\": 0.25, \"recall\": 0.15384615384615385, \"f1-score\": 0.1904761904761905, \"support\": 13.0}, \"request\": {\"precision\": 0.8181818181818182, \"recall\": 0.5625, \"f1-score\": 0.6666666666666666, \"support\": 16.0}, \"accuracy\": 0.4888888888888889, \"macro avg\": {\"precision\": 0.5126601199347101, \"recall\": 0.39449179292929293, \"f1-score\": 0.4269509445626284, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.47632523517769415, \"recall\": 0.4888888888888889, \"f1-score\": 0.46266093726956653, \"support\": 135.0}}", "eval_f1_macro": 42.69509445626284, "eval_f1_micro": 48.888888888888886, "eval_f1_weighted": 46.26609372695665, "eval_loss": 0.5566144585609436, "eval_runtime": 2.1877, "eval_samples_per_second": 61.709, "step": 136 }, { "epoch": 9.0, "learning_rate": 0.000135, "loss": 0.6661, "step": 153 }, { "epoch": 9.0, "eval_accuracy": 54.074074074074076, "eval_average_metrics": 54.03755296584174, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6, \"recall\": 0.3333333333333333, \"f1-score\": 0.42857142857142855, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7777777777777778, \"recall\": 0.6363636363636364, \"f1-score\": 0.7000000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.5581395348837209, \"recall\": 0.6666666666666666, \"f1-score\": 0.6075949367088608, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7272727272727273, \"recall\": 0.41025641025641024, \"f1-score\": 0.5245901639344261, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.3333333333333333, \"f1-score\": 0.46153846153846156, \"support\": 9.0}, \"rejection\": {\"precision\": 0.2727272727272727, \"recall\": 0.46153846153846156, \"f1-score\": 0.3428571428571428, \"support\": 13.0}, \"request\": {\"precision\": 0.4482758620689655, \"recall\": 0.8125, \"f1-score\": 0.5777777777777777, \"support\": 16.0}, \"accuracy\": 0.5407407407407407, \"macro avg\": {\"precision\": 0.6417741468413081, \"recall\": 0.5192489801864801, \"f1-score\": 0.5386995722568455, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.60651913713751, \"recall\": 0.5407407407407407, \"f1-score\": 0.5413210648953426, \"support\": 135.0}}", "eval_f1_macro": 53.86995722568455, "eval_f1_micro": 54.074074074074076, "eval_f1_weighted": 54.13210648953427, "eval_loss": 0.511962890625, "eval_runtime": 2.1819, "eval_samples_per_second": 61.874, "step": 153 }, { "epoch": 10.0, "learning_rate": 0.00015, "loss": 0.6131, "step": 170 }, { "epoch": 10.0, "eval_accuracy": 57.03703703703704, "eval_average_metrics": 57.547429298101335, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6111111111111112, \"recall\": 0.6111111111111112, \"f1-score\": 0.6111111111111112, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7391304347826086, \"recall\": 0.4358974358974359, \"f1-score\": 0.5483870967741935, \"support\": 39.0}, \"question\": {\"precision\": 0.8, \"recall\": 0.4444444444444444, \"f1-score\": 0.5714285714285714, \"support\": 9.0}, \"rejection\": {\"precision\": 0.3157894736842105, \"recall\": 0.46153846153846156, \"f1-score\": 0.37499999999999994, \"support\": 13.0}, \"request\": {\"precision\": 0.42424242424242425, \"recall\": 0.875, \"f1-score\": 0.5714285714285714, \"support\": 16.0}, \"accuracy\": 0.5703703703703704, \"macro avg\": {\"precision\": 0.6664789856723495, \"recall\": 0.576352466977467, \"f1-score\": 0.5870368430852302, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6322059283767903, \"recall\": 0.5703703703703704, \"f1-score\": 0.5741195880980826, \"support\": 135.0}}", "eval_f1_macro": 58.703684308523016, "eval_f1_micro": 57.03703703703704, "eval_f1_weighted": 57.41195880980826, "eval_loss": 0.4787319600582123, "eval_runtime": 2.2206, "eval_samples_per_second": 60.794, "step": 170 }, { "epoch": 11.0, "learning_rate": 0.000165, "loss": 0.5787, "step": 187 }, { "epoch": 11.0, "eval_accuracy": 64.44444444444444, "eval_average_metrics": 65.62118906429522, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6333333333333333, \"recall\": 0.5277777777777778, \"f1-score\": 0.5757575757575758, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.5833333333333334, \"recall\": 0.717948717948718, \"f1-score\": 0.6436781609195402, \"support\": 39.0}, \"question\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"rejection\": {\"precision\": 0.7, \"recall\": 0.5384615384615384, \"f1-score\": 0.608695652173913, \"support\": 13.0}, \"request\": {\"precision\": 0.7333333333333333, \"recall\": 0.6875, \"f1-score\": 0.7096774193548386, \"support\": 16.0}, \"accuracy\": 0.6444444444444445, \"macro avg\": {\"precision\": 0.7084821428571428, \"recall\": 0.6890115093240092, \"f1-score\": 0.6939908069080865, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6505908289241622, \"recall\": 0.6444444444444445, \"f1-score\": 0.6419678667748333, \"support\": 135.0}}", "eval_f1_macro": 69.39908069080865, "eval_f1_micro": 64.44444444444444, "eval_f1_weighted": 64.19678667748333, "eval_loss": 0.4315463900566101, "eval_runtime": 2.336, "eval_samples_per_second": 57.791, "step": 187 }, { "epoch": 12.0, "learning_rate": 0.00017999999999999998, "loss": 0.5394, "step": 204 }, { "epoch": 12.0, "eval_accuracy": 68.14814814814815, "eval_average_metrics": 69.08949458669183, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.6153846153846154, \"recall\": 0.8888888888888888, \"f1-score\": 0.7272727272727274, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6, \"recall\": 0.8181818181818182, \"f1-score\": 0.6923076923076923, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6444444444444445, \"recall\": 0.8055555555555556, \"f1-score\": 0.7160493827160493, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6896551724137931, \"recall\": 0.5128205128205128, \"f1-score\": 0.5882352941176471, \"support\": 39.0}, \"question\": {\"precision\": 0.8333333333333334, \"recall\": 0.5555555555555556, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 13.0}, \"request\": {\"precision\": 0.7333333333333333, \"recall\": 0.6875, \"f1-score\": 0.7096774193548386, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7395188623636899, \"recall\": 0.7354858682983683, \"f1-score\": 0.7244826695435831, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6953210858957987, \"recall\": 0.6814814814814815, \"f1-score\": 0.6761341509611271, \"support\": 135.0}}", "eval_f1_macro": 72.4482669543583, "eval_f1_micro": 68.14814814814815, "eval_f1_weighted": 67.6134150961127, "eval_loss": 0.3929964601993561, "eval_runtime": 2.2653, "eval_samples_per_second": 59.595, "step": 204 }, { "epoch": 13.0, "learning_rate": 0.000195, "loss": 0.4445, "step": 221 }, { "epoch": 13.0, "eval_accuracy": 68.14814814814815, "eval_average_metrics": 69.54531108529271, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6136363636363636, \"recall\": 0.75, \"f1-score\": 0.6749999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.631578947368421, \"recall\": 0.6153846153846154, \"f1-score\": 0.6233766233766234, \"support\": 39.0}, \"question\": {\"precision\": 0.8571428571428571, \"recall\": 0.6666666666666666, \"f1-score\": 0.75, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8571428571428571, \"recall\": 0.46153846153846156, \"f1-score\": 0.6, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7767885896997739, \"recall\": 0.7221311674436675, \"f1-score\": 0.7386757543926662, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6968149297096665, \"recall\": 0.6814814814814815, \"f1-score\": 0.6801737260560791, \"support\": 135.0}}", "eval_f1_macro": 73.86757543926662, "eval_f1_micro": 68.14814814814815, "eval_f1_weighted": 68.0173726056079, "eval_loss": 0.3637169301509857, "eval_runtime": 2.1262, "eval_samples_per_second": 63.494, "step": 221 }, { "epoch": 14.0, "learning_rate": 0.00020999999999999998, "loss": 0.4284, "step": 238 }, { "epoch": 14.0, "eval_accuracy": 68.88888888888889, "eval_average_metrics": 70.04225470037278, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.6363636363636364, \"f1-score\": 0.6666666666666666, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6511627906976745, \"recall\": 0.7777777777777778, \"f1-score\": 0.7088607594936709, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5384615384615384, \"f1-score\": 0.6176470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9, \"recall\": 0.6923076923076923, \"f1-score\": 0.7826086956521738, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7628809968099262, \"recall\": 0.7316554972804973, \"f1-score\": 0.7363724073834267, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.711127376440127, \"recall\": 0.6888888888888889, \"f1-score\": 0.6875400028537069, \"support\": 135.0}}", "eval_f1_macro": 73.63724073834267, "eval_f1_micro": 68.88888888888889, "eval_f1_weighted": 68.7540002853707, "eval_loss": 0.35276439785957336, "eval_runtime": 2.1486, "eval_samples_per_second": 62.832, "step": 238 }, { "epoch": 15.0, "learning_rate": 0.000225, "loss": 0.4125, "step": 255 }, { "epoch": 15.0, "eval_accuracy": 66.66666666666666, "eval_average_metrics": 67.82947247449044, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.5918367346938775, \"recall\": 0.8055555555555556, \"f1-score\": 0.6823529411764706, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8571428571428571, \"recall\": 0.3076923076923077, \"f1-score\": 0.45283018867924535, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6153846153846154, \"f1-score\": 0.761904761904762, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.6666666666666666, \"macro avg\": {\"precision\": 0.761308929166072, \"recall\": 0.7724237567987569, \"f1-score\": 0.7343863238519761, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7304495152114199, \"recall\": 0.6666666666666666, \"f1-score\": 0.6454592417943081, \"support\": 135.0}}", "eval_f1_macro": 73.43863238519761, "eval_f1_micro": 66.66666666666666, "eval_f1_weighted": 64.54592417943081, "eval_loss": 0.36039280891418457, "eval_runtime": 2.0664, "eval_samples_per_second": 65.332, "step": 255 }, { "epoch": 16.0, "learning_rate": 0.00023999999999999998, "loss": 0.3537, "step": 272 }, { "epoch": 16.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.72736523652416, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6521739130434783, \"recall\": 0.8333333333333334, \"f1-score\": 0.7317073170731708, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8095238095238095, \"recall\": 0.4358974358974359, \"f1-score\": 0.5666666666666667, \"support\": 39.0}, \"question\": {\"precision\": 0.8333333333333334, \"recall\": 0.5555555555555556, \"f1-score\": 0.6666666666666667, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6923076923076923, \"f1-score\": 0.8181818181818181, \"support\": 13.0}, \"request\": {\"precision\": 0.5185185185185185, \"recall\": 0.875, \"f1-score\": 0.6511627906976744, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.8048186968023925, \"recall\": 0.7737592268842269, \"f1-score\": 0.7645378088268715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7534887832391858, \"recall\": 0.7037037037037037, \"f1-score\": 0.6971493932266877, \"support\": 135.0}}", "eval_f1_macro": 76.45378088268716, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 69.71493932266877, "eval_loss": 0.3415175974369049, "eval_runtime": 2.2011, "eval_samples_per_second": 61.334, "step": 272 }, { "epoch": 17.0, "learning_rate": 0.00025499999999999996, "loss": 0.3307, "step": 289 }, { "epoch": 17.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.80907924352218, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7720217737121258, \"recall\": 0.7883704836829837, \"f1-score\": 0.7742597755505364, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7391776099879956, \"recall\": 0.725925925925926, \"f1-score\": 0.7262515423384989, \"support\": 135.0}}", "eval_f1_macro": 77.42597755505363, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.62515423384988, "eval_loss": 0.320279061794281, "eval_runtime": 2.1547, "eval_samples_per_second": 62.654, "step": 289 }, { "epoch": 18.0, "learning_rate": 0.00027, "loss": 0.2829, "step": 306 }, { "epoch": 18.0, "eval_accuracy": 75.55555555555556, "eval_average_metrics": 76.66401588989308, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6829268292682927, \"recall\": 0.7777777777777778, \"f1-score\": 0.7272727272727273, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.7857142857142857, \"recall\": 0.6875, \"f1-score\": 0.7333333333333334, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.7981489039415868, \"recall\": 0.807868346930847, \"f1-score\": 0.7989500265816055, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7661238851482753, \"recall\": 0.7555555555555555, \"f1-score\": 0.7564994979030066, \"support\": 135.0}}", "eval_f1_macro": 79.89500265816055, "eval_f1_micro": 75.55555555555556, "eval_f1_weighted": 75.64994979030067, "eval_loss": 0.3107147812843323, "eval_runtime": 2.1712, "eval_samples_per_second": 62.178, "step": 306 }, { "epoch": 19.0, "learning_rate": 0.000285, "loss": 0.27, "step": 323 }, { "epoch": 19.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.65875801590278, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6078431372549019, \"recall\": 0.8611111111111112, \"f1-score\": 0.7126436781609194, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8823529411764706, \"recall\": 0.38461538461538464, \"f1-score\": 0.5357142857142858, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7684180606974724, \"recall\": 0.7708758255633257, \"f1-score\": 0.7515478083829743, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7415855386443622, \"recall\": 0.6962962962962963, \"f1-score\": 0.6822099196605441, \"support\": 135.0}}", "eval_f1_macro": 75.15478083829744, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 68.22099196605443, "eval_loss": 0.3932558298110962, "eval_runtime": 2.1067, "eval_samples_per_second": 64.082, "step": 323 }, { "epoch": 20.0, "learning_rate": 0.0003, "loss": 0.2814, "step": 340 }, { "epoch": 20.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.63929543961528, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6904761904761905, \"recall\": 0.7435897435897436, \"f1-score\": 0.7160493827160495, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6875, \"recall\": 0.6875, \"f1-score\": 0.6875, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7865632631257631, \"recall\": 0.731479458041958, \"f1-score\": 0.7439023969088441, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7466218966218966, \"recall\": 0.7333333333333333, \"f1-score\": 0.7350027540091005, \"support\": 135.0}}", "eval_f1_macro": 74.39023969088441, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.50027540091006, "eval_loss": 0.35566049814224243, "eval_runtime": 2.1665, "eval_samples_per_second": 62.312, "step": 340 }, { "epoch": 21.0, "learning_rate": 0.00029833333333333334, "loss": 0.242, "step": 357 }, { "epoch": 21.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.95524042950514, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7941800524037366, \"recall\": 0.8099504662004662, \"f1-score\": 0.7964169000933707, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7463545226703122, \"recall\": 0.7333333333333333, \"f1-score\": 0.7351260504201681, \"support\": 135.0}}", "eval_f1_macro": 79.64169000933707, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.51260504201682, "eval_loss": 0.33881324529647827, "eval_runtime": 2.177, "eval_samples_per_second": 62.012, "step": 357 }, { "epoch": 22.0, "learning_rate": 0.00029666666666666665, "loss": 0.1889, "step": 374 }, { "epoch": 22.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.62607689125785, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6511627906976745, \"recall\": 0.7777777777777778, \"f1-score\": 0.7088607594936709, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7855620155038759, \"recall\": 0.8003350815850816, \"f1-score\": 0.7871122658545654, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7467298306057996, \"recall\": 0.7333333333333333, \"f1-score\": 0.7312641431290826, \"support\": 135.0}}", "eval_f1_macro": 78.71122658545654, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.12641431290825, "eval_loss": 0.38223153352737427, "eval_runtime": 2.1885, "eval_samples_per_second": 61.685, "step": 374 }, { "epoch": 23.0, "learning_rate": 0.00029499999999999996, "loss": 0.1787, "step": 391 }, { "epoch": 23.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.05134386837818, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7241379310344828, \"recall\": 0.5384615384615384, \"f1-score\": 0.6176470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7613135503335141, \"recall\": 0.7901065947940948, \"f1-score\": 0.7691766030736619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7284590496326734, \"recall\": 0.7185185185185186, \"f1-score\": 0.7158401146244283, \"support\": 135.0}}", "eval_f1_macro": 76.91766030736619, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 71.58401146244283, "eval_loss": 0.40378537774086, "eval_runtime": 2.2888, "eval_samples_per_second": 58.982, "step": 391 }, { "epoch": 24.0, "learning_rate": 0.00029333333333333327, "loss": 0.1458, "step": 408 }, { "epoch": 24.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.88129852922421, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.5128205128205128, \"f1-score\": 0.625, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.75, \"recall\": 0.9230769230769231, \"f1-score\": 0.8275862068965517, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7539396367521367, \"recall\": 0.8122207653457654, \"f1-score\": 0.7701591825594307, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7255460588793923, \"recall\": 0.7037037037037037, \"f1-score\": 0.6976853512021299, \"support\": 135.0}}", "eval_f1_macro": 77.01591825594308, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 69.768535120213, "eval_loss": 0.4484809637069702, "eval_runtime": 2.2039, "eval_samples_per_second": 61.256, "step": 408 }, { "epoch": 25.0, "learning_rate": 0.00029166666666666664, "loss": 0.1755, "step": 425 }, { "epoch": 25.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.30296325475025, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6595744680851063, \"recall\": 0.8611111111111112, \"f1-score\": 0.7469879518072289, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8461538461538461, \"recall\": 0.5641025641025641, \"f1-score\": 0.676923076923077, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6666666666666666, \"recall\": 0.75, \"f1-score\": 0.7058823529411765, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7791251301889599, \"recall\": 0.7255244755244756, \"f1-score\": 0.7344306944480657, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7577047066408767, \"recall\": 0.7333333333333333, \"f1-score\": 0.7310211690752781, \"support\": 135.0}}", "eval_f1_macro": 73.44306944480657, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.1021169075278, "eval_loss": 0.47192898392677307, "eval_runtime": 2.1604, "eval_samples_per_second": 62.489, "step": 425 }, { "epoch": 26.0, "learning_rate": 0.00029, "loss": 0.1193, "step": 442 }, { "epoch": 26.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.94665416513004, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6578947368421053, \"recall\": 0.6944444444444444, \"f1-score\": 0.6756756756756757, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8037195260879472, \"recall\": 0.7995337995337994, \"f1-score\": 0.7956619448395763, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7480504291030607, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355375550989586, \"support\": 135.0}}", "eval_f1_macro": 79.56619448395764, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.55375550989585, "eval_loss": 0.45392414927482605, "eval_runtime": 2.2435, "eval_samples_per_second": 60.174, "step": 442 }, { "epoch": 27.0, "learning_rate": 0.0002883333333333333, "loss": 0.1263, "step": 459 }, { "epoch": 27.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 74.0950271183071, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65, \"recall\": 0.7222222222222222, \"f1-score\": 0.6842105263157895, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7858405483405484, \"recall\": 0.7965957653457654, \"f1-score\": 0.7843223575106921, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7482972582972582, \"recall\": 0.725925925925926, \"f1-score\": 0.7276268753697399, \"support\": 135.0}}", "eval_f1_macro": 78.43223575106921, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.762687536974, "eval_loss": 0.46841171383857727, "eval_runtime": 2.1169, "eval_samples_per_second": 63.773, "step": 459 }, { "epoch": 28.0, "learning_rate": 0.0002866666666666667, "loss": 0.1043, "step": 476 }, { "epoch": 28.0, "eval_accuracy": 68.88888888888889, "eval_average_metrics": 69.42866207163016, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8333333333333334, \"recall\": 0.5128205128205128, \"f1-score\": 0.6349206349206349, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7394724025974025, \"recall\": 0.7357651029526029, \"f1-score\": 0.7130452134668899, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7225299021965689, \"recall\": 0.6888888888888889, \"f1-score\": 0.6863234916205385, \"support\": 135.0}}", "eval_f1_macro": 71.30452134668899, "eval_f1_micro": 68.88888888888889, "eval_f1_weighted": 68.63234916205386, "eval_loss": 0.49356982111930847, "eval_runtime": 2.0925, "eval_samples_per_second": 64.518, "step": 476 }, { "epoch": 29.0, "learning_rate": 0.000285, "loss": 0.1174, "step": 493 }, { "epoch": 29.0, "eval_accuracy": 68.14814814814815, "eval_average_metrics": 69.89790682417316, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6206896551724138, \"recall\": 0.5, \"f1-score\": 0.5538461538461539, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.5454545454545454, \"recall\": 0.6666666666666666, \"f1-score\": 0.6, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.4827586206896552, \"recall\": 0.875, \"f1-score\": 0.6222222222222222, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7515477604986782, \"recall\": 0.7737592268842268, \"f1-score\": 0.7502473425267543, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7103981308931253, \"recall\": 0.6814814814814815, \"f1-score\": 0.6827059674772094, \"support\": 135.0}}", "eval_f1_macro": 75.02473425267543, "eval_f1_micro": 68.14814814814815, "eval_f1_weighted": 68.27059674772093, "eval_loss": 0.5587795972824097, "eval_runtime": 2.2187, "eval_samples_per_second": 60.847, "step": 493 }, { "epoch": 30.0, "learning_rate": 0.0002833333333333333, "loss": 0.0689, "step": 510 }, { "epoch": 30.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.94861619800345, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7931034482758621, \"recall\": 0.5897435897435898, \"f1-score\": 0.676470588235294, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7848776232528335, \"recall\": 0.7861790986790986, \"f1-score\": 0.7791600854100853, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7445433448587359, \"recall\": 0.725925925925926, \"f1-score\": 0.7269327106582008, \"support\": 135.0}}", "eval_f1_macro": 77.91600854100852, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.69327106582007, "eval_loss": 0.539673388004303, "eval_runtime": 2.1316, "eval_samples_per_second": 63.333, "step": 510 }, { "epoch": 31.0, "learning_rate": 0.00028166666666666666, "loss": 0.0539, "step": 527 }, { "epoch": 31.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.33781120741372, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7096774193548387, \"recall\": 0.5641025641025641, \"f1-score\": 0.6285714285714286, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7561203917050692, \"recall\": 0.7387031371406372, \"f1-score\": 0.7282430580928023, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7082807077430734, \"recall\": 0.6962962962962963, \"f1-score\": 0.6926767976111539, \"support\": 135.0}}", "eval_f1_macro": 72.82430580928023, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.26767976111539, "eval_loss": 0.5683187246322632, "eval_runtime": 2.1767, "eval_samples_per_second": 62.021, "step": 527 }, { "epoch": 32.0, "learning_rate": 0.00028, "loss": 0.0591, "step": 544 }, { "epoch": 32.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.58032751159402, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7027027027027027, \"recall\": 0.7222222222222222, \"f1-score\": 0.7123287671232876, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7916666666666666, \"recall\": 0.48717948717948717, \"f1-score\": 0.6031746031746031, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.738858691983692, \"recall\": 0.7976641414141414, \"f1-score\": 0.757487853810853, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7226086476086476, \"recall\": 0.7037037037037037, \"f1-score\": 0.6983178392455008, \"support\": 135.0}}", "eval_f1_macro": 75.7487853810853, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 69.83178392455008, "eval_loss": 0.6107626557350159, "eval_runtime": 2.1901, "eval_samples_per_second": 61.641, "step": 544 }, { "epoch": 33.0, "learning_rate": 0.00027833333333333334, "loss": 0.0539, "step": 561 }, { "epoch": 33.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.67097529451648, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7804387991887991, \"recall\": 0.7399718337218337, \"f1-score\": 0.7468834223239258, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7408524975191642, \"recall\": 0.7333333333333333, \"f1-score\": 0.7332889227900669, \"support\": 135.0}}", "eval_f1_macro": 74.68834223239257, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.32889227900668, "eval_loss": 0.5886625051498413, "eval_runtime": 2.2726, "eval_samples_per_second": 59.404, "step": 561 }, { "epoch": 34.0, "learning_rate": 0.00027666666666666665, "loss": 0.0734, "step": 578 }, { "epoch": 34.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 75.0892478093565, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7929446284709443, \"recall\": 0.7491744366744366, \"f1-score\": 0.757480864546082, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7589443694706852, \"recall\": 0.7481481481481481, \"f1-score\": 0.749792751531882, \"support\": 135.0}}", "eval_f1_macro": 75.7480864546082, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 74.9792751531882, "eval_loss": 0.5679999589920044, "eval_runtime": 2.2089, "eval_samples_per_second": 61.116, "step": 578 }, { "epoch": 35.0, "learning_rate": 0.00027499999999999996, "loss": 0.0343, "step": 595 }, { "epoch": 35.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.3823414203499, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6585365853658537, \"recall\": 0.75, \"f1-score\": 0.7012987012987012, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7803482087018672, \"recall\": 0.7343628593628594, \"f1-score\": 0.7409159272416515, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7359272164150212, \"recall\": 0.7185185185185186, \"f1-score\": 0.7173406925353076, \"support\": 135.0}}", "eval_f1_macro": 74.09159272416514, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 71.73406925353078, "eval_loss": 0.64279705286026, "eval_runtime": 2.2258, "eval_samples_per_second": 60.654, "step": 595 }, { "epoch": 36.0, "learning_rate": 0.00027333333333333333, "loss": 0.0256, "step": 612 }, { "epoch": 36.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.27192027743499, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6470588235294118, \"recall\": 0.6111111111111112, \"f1-score\": 0.6285714285714287, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.774467344387318, \"recall\": 0.7373008935508936, \"f1-score\": 0.7363179712444419, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7299169542548802, \"recall\": 0.7037037037037037, \"f1-score\": 0.7071514324455502, \"support\": 135.0}}", "eval_f1_macro": 73.63179712444419, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.71514324455502, "eval_loss": 0.6609192490577698, "eval_runtime": 2.2472, "eval_samples_per_second": 60.074, "step": 612 }, { "epoch": 37.0, "learning_rate": 0.00027166666666666664, "loss": 0.0369, "step": 629 }, { "epoch": 37.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.77168024471725, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7861613248710023, \"recall\": 0.8058651418026419, \"f1-score\": 0.7918380042511899, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7395884688357807, \"recall\": 0.7333333333333333, \"f1-score\": 0.7323625388708334, \"support\": 135.0}}", "eval_f1_macro": 79.18380042511899, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.23625388708334, "eval_loss": 0.6587929725646973, "eval_runtime": 2.1685, "eval_samples_per_second": 62.254, "step": 629 }, { "epoch": 38.0, "learning_rate": 0.00027, "loss": 0.0411, "step": 646 }, { "epoch": 38.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.4984910288592, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7916610299422799, \"recall\": 0.7314915986790986, \"f1-score\": 0.741379475703325, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7429092512425847, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215231284140066, \"support\": 135.0}}", "eval_f1_macro": 74.13794757033249, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.15231284140066, "eval_loss": 0.717117190361023, "eval_runtime": 2.2579, "eval_samples_per_second": 59.789, "step": 646 }, { "epoch": 39.0, "learning_rate": 0.0002683333333333333, "loss": 0.0188, "step": 663 }, { "epoch": 39.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 75.21675953993977, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7982278836574502, \"recall\": 0.7576000388500388, \"f1-score\": 0.7635135515393372, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7587398134869756, \"recall\": 0.7481481481481481, \"f1-score\": 0.7488605337619573, \"support\": 135.0}}", "eval_f1_macro": 76.35135515393372, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 74.88605337619573, "eval_loss": 0.6625694036483765, "eval_runtime": 2.1219, "eval_samples_per_second": 63.622, "step": 663 }, { "epoch": 40.0, "learning_rate": 0.0002666666666666666, "loss": 0.0232, "step": 680 }, { "epoch": 40.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.41849073937567, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7912581597759617, \"recall\": 0.8052641802641802, \"f1-score\": 0.7945949432404541, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7460020576119648, \"recall\": 0.7407407407407407, \"f1-score\": 0.7406632048530915, \"support\": 135.0}}", "eval_f1_macro": 79.45949432404542, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.06632048530913, "eval_loss": 0.7407746315002441, "eval_runtime": 2.1843, "eval_samples_per_second": 61.805, "step": 680 }, { "epoch": 41.0, "learning_rate": 0.000265, "loss": 0.034, "step": 697 }, { "epoch": 41.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 72.10801080339358, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6285714285714286, \"recall\": 0.6111111111111112, \"f1-score\": 0.619718309859155, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7691468253968254, \"recall\": 0.7905193764568765, \"f1-score\": 0.7703084495012319, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327513227513227, \"recall\": 0.7037037037037037, \"f1-score\": 0.7066045752271038, \"support\": 135.0}}", "eval_f1_macro": 77.0308449501232, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.66045752271039, "eval_loss": 0.8014240264892578, "eval_runtime": 2.1576, "eval_samples_per_second": 62.569, "step": 697 }, { "epoch": 42.0, "learning_rate": 0.0002633333333333333, "loss": 0.0393, "step": 714 }, { "epoch": 42.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 72.91137239666652, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6428571428571429, \"recall\": 0.75, \"f1-score\": 0.6923076923076924, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7874594155844157, \"recall\": 0.7123154623154623, \"f1-score\": 0.7361968097262215, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.741950456950457, \"recall\": 0.725925925925926, \"f1-score\": 0.7284062342885872, \"support\": 135.0}}", "eval_f1_macro": 73.61968097262215, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.84062342885873, "eval_loss": 0.7516341805458069, "eval_runtime": 2.2389, "eval_samples_per_second": 60.297, "step": 714 }, { "epoch": 43.0, "learning_rate": 0.00026166666666666667, "loss": 0.0187, "step": 731 }, { "epoch": 43.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 71.6853552122398, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6363636363636364, \"recall\": 0.5833333333333334, \"f1-score\": 0.6086956521739131, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7419354838709677, \"recall\": 0.5897435897435898, \"f1-score\": 0.6571428571428573, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.784555908547844, \"recall\": 0.7870471542346542, \"f1-score\": 0.7754307414417895, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7231971245710792, \"recall\": 0.6962962962962963, \"f1-score\": 0.6993908744552095, \"support\": 135.0}}", "eval_f1_macro": 77.54307414417895, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.93908744552097, "eval_loss": 0.8274635672569275, "eval_runtime": 2.2726, "eval_samples_per_second": 59.402, "step": 731 }, { "epoch": 44.0, "learning_rate": 0.00026, "loss": 0.0326, "step": 748 }, { "epoch": 44.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.29350182491167, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.778793688949939, \"recall\": 0.7780205905205906, \"f1-score\": 0.7740684627673118, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7306166056166056, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206345731921179, \"support\": 135.0}}", "eval_f1_macro": 77.40684627673117, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06345731921179, "eval_loss": 0.7537038326263428, "eval_runtime": 2.251, "eval_samples_per_second": 59.972, "step": 748 }, { "epoch": 45.0, "learning_rate": 0.00025833333333333334, "loss": 0.0446, "step": 765 }, { "epoch": 45.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.55936282460405, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7762371615312792, \"recall\": 0.7913752913752914, \"f1-score\": 0.7800986983223825, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7445440398381575, \"recall\": 0.7333333333333333, \"f1-score\": 0.735609147995113, \"support\": 135.0}}", "eval_f1_macro": 78.00986983223825, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.5609147995113, "eval_loss": 0.7628147006034851, "eval_runtime": 2.1831, "eval_samples_per_second": 61.838, "step": 765 }, { "epoch": 46.0, "learning_rate": 0.00025666666666666665, "loss": 0.0222, "step": 782 }, { "epoch": 46.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.61743270582245, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7865956727518593, \"recall\": 0.8073462995337995, \"f1-score\": 0.7881557617778407, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7388550822627902, \"recall\": 0.7185185185185186, \"f1-score\": 0.71950450941802, \"support\": 135.0}}", "eval_f1_macro": 78.81557617778408, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 71.950450941802, "eval_loss": 0.8409448266029358, "eval_runtime": 2.2893, "eval_samples_per_second": 58.969, "step": 782 }, { "epoch": 47.0, "learning_rate": 0.00025499999999999996, "loss": 0.0137, "step": 799 }, { "epoch": 47.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 72.17480498243586, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8064516129032258, \"recall\": 0.6410256410256411, \"f1-score\": 0.7142857142857142, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7435458256022771, \"recall\": 0.7664687742812744, \"f1-score\": 0.7519798629432732, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7228986803180352, \"recall\": 0.7111111111111111, \"f1-score\": 0.712790114131939, \"support\": 135.0}}", "eval_f1_macro": 75.19798629432732, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.27901141319391, "eval_loss": 0.8047698140144348, "eval_runtime": 2.1643, "eval_samples_per_second": 62.375, "step": 799 }, { "epoch": 48.0, "learning_rate": 0.00025333333333333333, "loss": 0.0209, "step": 816 }, { "epoch": 48.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.5841574785611, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.48, \"recall\": 0.75, \"f1-score\": 0.5853658536585366, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.78078081232493, \"recall\": 0.8030060217560218, \"f1-score\": 0.7854538367301481, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7372613341632949, \"recall\": 0.7185185185185186, \"f1-score\": 0.7208754253752583, \"support\": 135.0}}", "eval_f1_macro": 78.54538367301481, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.08754253752583, "eval_loss": 0.8688657283782959, "eval_runtime": 2.1847, "eval_samples_per_second": 61.794, "step": 816 }, { "epoch": 49.0, "learning_rate": 0.00025166666666666664, "loss": 0.0091, "step": 833 }, { "epoch": 49.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.75127888430008, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8125, \"recall\": 0.6666666666666666, \"f1-score\": 0.7323943661971831, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7851934523809524, \"recall\": 0.7992667055167055, \"f1-score\": 0.7875581374336444, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.748179012345679, \"recall\": 0.7333333333333333, \"f1-score\": 0.735826351271692, \"support\": 135.0}}", "eval_f1_macro": 78.75581374336443, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.58263512716921, "eval_loss": 0.8018043637275696, "eval_runtime": 2.2936, "eval_samples_per_second": 58.859, "step": 833 }, { "epoch": 50.0, "learning_rate": 0.00025, "loss": 0.0097, "step": 850 }, { "epoch": 50.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.57898288245715, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65, \"recall\": 0.7222222222222222, \"f1-score\": 0.6842105263157895, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8387096774193549, \"recall\": 0.6666666666666666, \"f1-score\": 0.7428571428571428, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8063538611925709, \"recall\": 0.7948475135975136, \"f1-score\": 0.795720639514387, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7636070381231672, \"recall\": 0.7407407407407407, \"f1-score\": 0.7459571943024179, \"support\": 135.0}}", "eval_f1_macro": 79.5720639514387, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.59571943024179, "eval_loss": 0.8517288565635681, "eval_runtime": 2.1756, "eval_samples_per_second": 62.051, "step": 850 }, { "epoch": 51.0, "learning_rate": 0.0002483333333333333, "loss": 0.0193, "step": 867 }, { "epoch": 51.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 72.8732174399158, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7058823529411765, \"recall\": 0.6153846153846154, \"f1-score\": 0.6575342465753424, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7702113042186571, \"recall\": 0.8102175602175602, \"f1-score\": 0.7806177067478437, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7336914900640391, \"recall\": 0.7111111111111111, \"f1-score\": 0.7120887686265667, \"support\": 135.0}}", "eval_f1_macro": 78.06177067478437, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.20887686265667, "eval_loss": 0.9536928534507751, "eval_runtime": 2.1272, "eval_samples_per_second": 63.463, "step": 867 }, { "epoch": 52.0, "learning_rate": 0.0002466666666666666, "loss": 0.0046, "step": 884 }, { "epoch": 52.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.64170538878341, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6829268292682927, \"recall\": 0.717948717948718, \"f1-score\": 0.7000000000000001, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7706033088556692, \"recall\": 0.7170563811188811, \"f1-score\": 0.7315823058619944, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7196237044807725, \"recall\": 0.7111111111111111, \"f1-score\": 0.7118636874671196, \"support\": 135.0}}", "eval_f1_macro": 73.15823058619944, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.18636874671196, "eval_loss": 0.8713467121124268, "eval_runtime": 2.2641, "eval_samples_per_second": 59.627, "step": 884 }, { "epoch": 53.0, "learning_rate": 0.000245, "loss": 0.0085, "step": 901 }, { "epoch": 53.0, "eval_accuracy": 68.88888888888889, "eval_average_metrics": 69.60698215446021, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.5277777777777778, \"f1-score\": 0.6031746031746033, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7484503734503735, \"recall\": 0.721930846930847, \"f1-score\": 0.7174509189640768, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7075614075614075, \"recall\": 0.6888888888888889, \"f1-score\": 0.6890505894365544, \"support\": 135.0}}", "eval_f1_macro": 71.74509189640769, "eval_f1_micro": 68.88888888888889, "eval_f1_weighted": 68.90505894365543, "eval_loss": 0.9699578285217285, "eval_runtime": 2.2097, "eval_samples_per_second": 61.095, "step": 901 }, { "epoch": 54.0, "learning_rate": 0.0002433333333333333, "loss": 0.0181, "step": 918 }, { "epoch": 54.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.51024557847745, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.875, \"recall\": 0.7777777777777778, \"f1-score\": 0.823529411764706, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7, \"recall\": 0.6363636363636364, \"f1-score\": 0.6666666666666666, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.717948717948718, \"f1-score\": 0.708860759493671, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7724702380952382, \"recall\": 0.7056138306138307, \"f1-score\": 0.7247131858217115, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7247113997113998, \"recall\": 0.7111111111111111, \"f1-score\": 0.7134744150951646, \"support\": 135.0}}", "eval_f1_macro": 72.47131858217115, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.34744150951646, "eval_loss": 0.9296635389328003, "eval_runtime": 2.1343, "eval_samples_per_second": 63.252, "step": 918 }, { "epoch": 55.0, "learning_rate": 0.00024166666666666664, "loss": 0.0055, "step": 935 }, { "epoch": 55.0, "eval_accuracy": 68.14814814814815, "eval_average_metrics": 69.88210719785789, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8181818181818182, \"recall\": 1.0, \"f1-score\": 0.9, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6451612903225806, \"recall\": 0.5555555555555556, \"f1-score\": 0.5970149253731343, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7857142857142857, \"recall\": 0.5641025641025641, \"f1-score\": 0.6567164179104477, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 0.7857142857142857, \"recall\": 0.8461538461538461, \"f1-score\": 0.8148148148148148, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6814814814814815, \"macro avg\": {\"precision\": 0.7426706358157971, \"recall\": 0.7828950563325563, \"f1-score\": 0.751917200243583, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7054854941951716, \"recall\": 0.6814814814814815, \"f1-score\": 0.6804041247077698, \"support\": 135.0}}", "eval_f1_macro": 75.1917200243583, "eval_f1_micro": 68.14814814814815, "eval_f1_weighted": 68.04041247077699, "eval_loss": 0.9763553142547607, "eval_runtime": 2.1762, "eval_samples_per_second": 62.035, "step": 935 }, { "epoch": 56.0, "learning_rate": 0.00023999999999999998, "loss": 0.0051, "step": 952 }, { "epoch": 56.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.26569664902999, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7671443391552086, \"recall\": 0.7989206973581974, \"f1-score\": 0.7732298136645963, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7413002188364506, \"recall\": 0.7185185185185186, \"f1-score\": 0.720361015259566, \"support\": 135.0}}", "eval_f1_macro": 77.32298136645963, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.0361015259566, "eval_loss": 0.9423373341560364, "eval_runtime": 2.238, "eval_samples_per_second": 60.322, "step": 952 }, { "epoch": 57.0, "learning_rate": 0.0002383333333333333, "loss": 0.0114, "step": 969 }, { "epoch": 57.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 72.53354760332923, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.6363636363636364, \"f1-score\": 0.6363636363636364, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7597564519879099, \"recall\": 0.7908835955710956, \"f1-score\": 0.7683601409092514, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7236771810173345, \"recall\": 0.7111111111111111, \"f1-score\": 0.7107595410016957, \"support\": 135.0}}", "eval_f1_macro": 76.83601409092513, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.07595410016957, "eval_loss": 0.9203009605407715, "eval_runtime": 2.1386, "eval_samples_per_second": 63.126, "step": 969 }, { "epoch": 58.0, "learning_rate": 0.00023666666666666663, "loss": 0.0072, "step": 986 }, { "epoch": 58.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.6642490769342, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"cancellation\": {\"precision\": 0.0, \"recall\": 0.0, \"f1-score\": 0.0, \"support\": 0.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7071727257211129, \"recall\": 0.7029213070879737, \"f1-score\": 0.6958756590822236, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7471616316777607, \"recall\": 0.7185185185185186, \"f1-score\": 0.7240017370132812, \"support\": 135.0}}", "eval_f1_macro": 78.28601164675014, "eval_f1_micro": 72.1189591078067, "eval_f1_weighted": 72.40017370132811, "eval_loss": 0.9572186470031738, "eval_runtime": 2.1451, "eval_samples_per_second": 62.935, "step": 986 }, { "epoch": 59.0, "learning_rate": 0.00023499999999999997, "loss": 0.0042, "step": 1003 }, { "epoch": 59.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.7649842113928, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6857142857142857, \"recall\": 0.6153846153846154, \"f1-score\": 0.6486486486486486, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7765705163233214, \"recall\": 0.7346967268842268, \"f1-score\": 0.736597489423982, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7294317448174391, \"recall\": 0.7111111111111111, \"f1-score\": 0.7117796568095075, \"support\": 135.0}}", "eval_f1_macro": 73.6597489423982, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.17796568095075, "eval_loss": 0.8926369547843933, "eval_runtime": 2.1674, "eval_samples_per_second": 62.286, "step": 1003 }, { "epoch": 60.0, "learning_rate": 0.0002333333333333333, "loss": 0.0108, "step": 1020 }, { "epoch": 60.0, "eval_accuracy": 67.4074074074074, "eval_average_metrics": 67.96942582571397, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.5555555555555556, \"f1-score\": 0.6153846153846154, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6923076923076923, \"f1-score\": 0.8181818181818181, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.674074074074074, \"macro avg\": {\"precision\": 0.7355337802536078, \"recall\": 0.6994160353535354, \"f1-score\": 0.6963058347068645, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.6961066081755738, \"recall\": 0.674074074074074, \"f1-score\": 0.674323050173546, \"support\": 135.0}}", "eval_f1_macro": 69.63058347068645, "eval_f1_micro": 67.4074074074074, "eval_f1_weighted": 67.4323050173546, "eval_loss": 0.9399789571762085, "eval_runtime": 2.1159, "eval_samples_per_second": 63.802, "step": 1020 }, { "epoch": 61.0, "learning_rate": 0.00023166666666666667, "loss": 0.0056, "step": 1037 }, { "epoch": 61.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 71.17152096147001, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.75, \"recall\": 1.0, \"f1-score\": 0.8571428571428571, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7037037037037037, \"recall\": 0.5277777777777778, \"f1-score\": 0.6031746031746033, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7443091168091167, \"recall\": 0.7997219794094794, \"f1-score\": 0.7591938053549183, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.719369420702754, \"recall\": 0.6962962962962963, \"f1-score\": 0.6950744405112893, \"support\": 135.0}}", "eval_f1_macro": 75.91938053549183, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.50744405112893, "eval_loss": 0.9959737062454224, "eval_runtime": 2.1462, "eval_samples_per_second": 62.902, "step": 1037 }, { "epoch": 62.0, "learning_rate": 0.00023, "loss": 0.0091, "step": 1054 }, { "epoch": 62.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.43021846765396, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7951003737641669, \"recall\": 0.7485856157731158, \"f1-score\": 0.754229181551163, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7622480329376882, \"recall\": 0.7407407407407407, \"f1-score\": 0.7414980756735141, \"support\": 135.0}}", "eval_f1_macro": 75.42291815511629, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.14980756735142, "eval_loss": 0.9423182010650635, "eval_runtime": 2.0725, "eval_samples_per_second": 65.137, "step": 1054 }, { "epoch": 63.0, "learning_rate": 0.0002283333333333333, "loss": 0.006, "step": 1071 }, { "epoch": 63.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.13630195174072, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6774193548387096, \"recall\": 0.5833333333333334, \"f1-score\": 0.626865671641791, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6, \"recall\": 0.6666666666666666, \"f1-score\": 0.631578947368421, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7705405145929339, \"recall\": 0.7309574106449107, \"f1-score\": 0.7339954289925987, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7191460431245379, \"recall\": 0.7037037037037037, \"f1-score\": 0.7040492416696227, \"support\": 135.0}}", "eval_f1_macro": 73.39954289925987, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.40492416696227, "eval_loss": 0.9775393009185791, "eval_runtime": 2.1426, "eval_samples_per_second": 63.009, "step": 1071 }, { "epoch": 64.0, "learning_rate": 0.00022666666666666663, "loss": 0.0071, "step": 1088 }, { "epoch": 64.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.50660628560995, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8333333333333334, \"recall\": 0.6410256410256411, \"f1-score\": 0.7246376811594204, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5185185185185185, \"recall\": 0.875, \"f1-score\": 0.6511627906976744, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7855794668294669, \"recall\": 0.7343507187257188, \"f1-score\": 0.7410246999652906, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7480406698925217, \"recall\": 0.7185185185185186, \"f1-score\": 0.7222025144220701, \"support\": 135.0}}", "eval_f1_macro": 74.10246999652907, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.22025144220703, "eval_loss": 1.0484141111373901, "eval_runtime": 2.2145, "eval_samples_per_second": 60.962, "step": 1088 }, { "epoch": 65.0, "learning_rate": 0.000225, "loss": 0.0034, "step": 1105 }, { "epoch": 65.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.9244200133918, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.9, \"recall\": 1.0, \"f1-score\": 0.9473684210526316, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.675, \"recall\": 0.6923076923076923, \"f1-score\": 0.6835443037974683, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5789473684210527, \"recall\": 0.6875, \"f1-score\": 0.6285714285714286, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7900560761394801, \"recall\": 0.7485734751359752, \"f1-score\": 0.7557596878082226, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7442556314967181, \"recall\": 0.7333333333333333, \"f1-score\": 0.7345504460607825, \"support\": 135.0}}", "eval_f1_macro": 75.57596878082227, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.45504460607825, "eval_loss": 0.943881094455719, "eval_runtime": 2.2664, "eval_samples_per_second": 59.565, "step": 1105 }, { "epoch": 66.0, "learning_rate": 0.00022333333333333333, "loss": 0.007, "step": 1122 }, { "epoch": 66.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.91255932928726, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.65625, \"recall\": 0.5833333333333334, \"f1-score\": 0.6176470588235293, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7927684294871795, \"recall\": 0.7234787781662781, \"f1-score\": 0.7393785014005602, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7358262108262108, \"recall\": 0.7111111111111111, \"f1-score\": 0.7149016495487084, \"support\": 135.0}}", "eval_f1_macro": 73.93785014005603, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.49016495487082, "eval_loss": 1.0599663257598877, "eval_runtime": 2.1786, "eval_samples_per_second": 61.965, "step": 1122 }, { "epoch": 67.0, "learning_rate": 0.00022166666666666667, "loss": 0.0042, "step": 1139 }, { "epoch": 67.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 75.12632131391055, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8064516129032258, \"recall\": 0.6410256410256411, \"f1-score\": 0.7142857142857142, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7984916367980883, \"recall\": 0.8247074106449106, \"f1-score\": 0.801042885612959, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7643603699278729, \"recall\": 0.7333333333333333, \"f1-score\": 0.7373433002767967, \"support\": 135.0}}", "eval_f1_macro": 80.1042885612959, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.73433002767968, "eval_loss": 1.0870040655136108, "eval_runtime": 2.253, "eval_samples_per_second": 59.919, "step": 1139 }, { "epoch": 68.0, "learning_rate": 0.00021999999999999995, "loss": 0.0048, "step": 1156 }, { "epoch": 68.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.55212011915016, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.791516354016354, \"recall\": 0.7966625388500388, \"f1-score\": 0.7820397286821705, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7515169329984145, \"recall\": 0.7185185185185186, \"f1-score\": 0.7230080390467987, \"support\": 135.0}}", "eval_f1_macro": 78.20397286821705, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.30080390467987, "eval_loss": 1.0407289266586304, "eval_runtime": 2.2127, "eval_samples_per_second": 61.011, "step": 1156 }, { "epoch": 69.0, "learning_rate": 0.0002183333333333333, "loss": 0.0026, "step": 1173 }, { "epoch": 69.0, "eval_accuracy": 68.88888888888889, "eval_average_metrics": 70.6530531645798, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.7777777777777778, \"f1-score\": 0.8750000000000001, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5, \"f1-score\": 0.5901639344262295, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7635660173160174, \"recall\": 0.783040743978244, \"f1-score\": 0.7556388197313891, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.731313772647106, \"recall\": 0.6888888888888889, \"f1-score\": 0.6927055290740252, \"support\": 135.0}}", "eval_f1_macro": 75.56388197313892, "eval_f1_micro": 68.88888888888889, "eval_f1_weighted": 69.27055290740252, "eval_loss": 1.0904052257537842, "eval_runtime": 2.1381, "eval_samples_per_second": 63.141, "step": 1173 }, { "epoch": 70.0, "learning_rate": 0.00021666666666666666, "loss": 0.0014, "step": 1190 }, { "epoch": 70.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.87504553531681, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7893565188172043, \"recall\": 0.814290743978244, \"f1-score\": 0.7918166361805148, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7587290919952211, \"recall\": 0.7333333333333333, \"f1-score\": 0.7365185185654917, \"support\": 135.0}}", "eval_f1_macro": 79.18166361805147, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.65185185654917, "eval_loss": 1.0420516729354858, "eval_runtime": 2.151, "eval_samples_per_second": 62.763, "step": 1190 }, { "epoch": 71.0, "learning_rate": 0.000215, "loss": 0.0047, "step": 1207 }, { "epoch": 71.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.96275924849695, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7944225059094784, \"recall\": 0.7552629662004662, \"f1-score\": 0.7562933597057457, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7554414742767257, \"recall\": 0.7333333333333333, \"f1-score\": 0.735550343567466, \"support\": 135.0}}", "eval_f1_macro": 75.62933597057457, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.5550343567466, "eval_loss": 1.0715558528900146, "eval_runtime": 2.2105, "eval_samples_per_second": 61.073, "step": 1207 }, { "epoch": 72.0, "learning_rate": 0.00021333333333333333, "loss": 0.0019, "step": 1224 }, { "epoch": 72.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.36672605045894, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7666666666666667, \"recall\": 0.6388888888888888, \"f1-score\": 0.696969696969697, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7698383520074696, \"recall\": 0.7404271076146076, \"f1-score\": 0.7363487615884876, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7427004876024484, \"recall\": 0.7185185185185186, \"f1-score\": 0.7212832433928325, \"support\": 135.0}}", "eval_f1_macro": 73.63487615884875, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.12832433928324, "eval_loss": 1.1170759201049805, "eval_runtime": 2.1545, "eval_samples_per_second": 62.659, "step": 1224 }, { "epoch": 73.0, "learning_rate": 0.00021166666666666667, "loss": 0.0009, "step": 1241 }, { "epoch": 73.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.45112685839807, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8214285714285714, \"recall\": 0.5897435897435898, \"f1-score\": 0.6865671641791046, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7725357975357976, \"recall\": 0.7965168512043512, \"f1-score\": 0.7786937025399319, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7492250958917624, \"recall\": 0.7333333333333333, \"f1-score\": 0.7326847051293242, \"support\": 135.0}}", "eval_f1_macro": 77.8693702539932, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.26847051293241, "eval_loss": 1.0454473495483398, "eval_runtime": 2.0945, "eval_samples_per_second": 64.453, "step": 1241 }, { "epoch": 74.0, "learning_rate": 0.00020999999999999998, "loss": 0.0137, "step": 1258 }, { "epoch": 74.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.70350070614597, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6363636363636364, \"recall\": 0.6363636363636364, \"f1-score\": 0.6363636363636364, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6097560975609756, \"recall\": 0.6944444444444444, \"f1-score\": 0.6493506493506495, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7840743537137566, \"recall\": 0.7153627622377623, \"f1-score\": 0.7315785053858942, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7402746783789677, \"recall\": 0.7111111111111111, \"f1-score\": 0.7143393006377224, \"support\": 135.0}}", "eval_f1_macro": 73.15785053858941, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.43393006377225, "eval_loss": 1.0791348218917847, "eval_runtime": 2.1939, "eval_samples_per_second": 61.535, "step": 1258 }, { "epoch": 75.0, "learning_rate": 0.00020833333333333332, "loss": 0.0254, "step": 1275 }, { "epoch": 75.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.53066318197898, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7937821165762342, \"recall\": 0.7505766802641802, \"f1-score\": 0.7575329088486984, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7541625694566871, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422121369489791, \"support\": 135.0}}", "eval_f1_macro": 75.75329088486984, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.22121369489791, "eval_loss": 0.9704490303993225, "eval_runtime": 2.225, "eval_samples_per_second": 60.674, "step": 1275 }, { "epoch": 76.0, "learning_rate": 0.00020666666666666666, "loss": 0.0039, "step": 1292 }, { "epoch": 76.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.88592559677284, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7785046518667209, \"recall\": 0.7366877913752914, \"f1-score\": 0.7404468801773039, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7311833743557882, \"recall\": 0.7111111111111111, \"f1-score\": 0.7127679214713875, \"support\": 135.0}}", "eval_f1_macro": 74.04468801773038, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.27679214713875, "eval_loss": 1.000522255897522, "eval_runtime": 2.2253, "eval_samples_per_second": 60.667, "step": 1292 }, { "epoch": 77.0, "learning_rate": 0.000205, "loss": 0.0095, "step": 1309 }, { "epoch": 77.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.50938740714625, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7787457160894661, \"recall\": 0.7404271076146076, \"f1-score\": 0.7417628359660171, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7426006493506493, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215756232827957, \"support\": 135.0}}", "eval_f1_macro": 74.17628359660172, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.15756232827957, "eval_loss": 0.9987503886222839, "eval_runtime": 2.0847, "eval_samples_per_second": 64.758, "step": 1309 }, { "epoch": 78.0, "learning_rate": 0.00020333333333333333, "loss": 0.0016, "step": 1326 }, { "epoch": 78.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.50535308482088, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7787042370711725, \"recall\": 0.7404271076146076, \"f1-score\": 0.74266339591752, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7378141889969847, \"recall\": 0.7185185185185186, \"f1-score\": 0.7205136904382781, \"support\": 135.0}}", "eval_f1_macro": 74.266339591752, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.0513690438278, "eval_loss": 1.0020543336868286, "eval_runtime": 2.1954, "eval_samples_per_second": 61.492, "step": 1326 }, { "epoch": 79.0, "learning_rate": 0.00020166666666666667, "loss": 0.0011, "step": 1343 }, { "epoch": 79.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.48732962316897, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8027706290849673, \"recall\": 0.751790743978244, \"f1-score\": 0.7583791127061329, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7516668482207698, \"recall\": 0.725925925925926, \"f1-score\": 0.7292622203687736, \"support\": 135.0}}", "eval_f1_macro": 75.8379112706133, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.92622203687736, "eval_loss": 1.041650652885437, "eval_runtime": 2.0975, "eval_samples_per_second": 64.363, "step": 1343 }, { "epoch": 80.0, "learning_rate": 0.00019999999999999998, "loss": 0.0022, "step": 1360 }, { "epoch": 80.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.38039917163168, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7955693164362518, \"recall\": 0.751790743978244, \"f1-score\": 0.7553531897837845, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7471616316777607, \"recall\": 0.725925925925926, \"f1-score\": 0.7280109252296305, \"support\": 135.0}}", "eval_f1_macro": 75.53531897837846, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.80109252296305, "eval_loss": 1.0494800806045532, "eval_runtime": 2.2088, "eval_samples_per_second": 61.119, "step": 1360 }, { "epoch": 81.0, "learning_rate": 0.00019833333333333332, "loss": 0.0126, "step": 1377 }, { "epoch": 81.0, "eval_accuracy": 68.88888888888889, "eval_average_metrics": 69.48276901714948, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.72, \"recall\": 0.5, \"f1-score\": 0.5901639344262295, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7443162393162392, \"recall\": 0.7262711247086246, \"f1-score\": 0.7127222322360148, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.716459639126306, \"recall\": 0.6888888888888889, \"f1-score\": 0.6888107506721864, \"support\": 135.0}}", "eval_f1_macro": 71.27222322360149, "eval_f1_micro": 68.88888888888889, "eval_f1_weighted": 68.88107506721865, "eval_loss": 1.0959280729293823, "eval_runtime": 2.2009, "eval_samples_per_second": 61.34, "step": 1377 }, { "epoch": 82.0, "learning_rate": 0.00019666666666666666, "loss": 0.0033, "step": 1394 }, { "epoch": 82.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.0222652316685, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6571428571428571, \"recall\": 0.6388888888888888, \"f1-score\": 0.647887323943662, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.717948717948718, \"f1-score\": 0.7567567567567569, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.52, \"recall\": 0.8125, \"f1-score\": 0.6341463414634146, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8054761904761905, \"recall\": 0.738678855866356, \"f1-score\": 0.7563368980085744, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7553862433862435, \"recall\": 0.7333333333333333, \"f1-score\": 0.737887044591499, \"support\": 135.0}}", "eval_f1_macro": 75.63368980085744, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.78870445914991, "eval_loss": 1.07533860206604, "eval_runtime": 2.33, "eval_samples_per_second": 57.939, "step": 1394 }, { "epoch": 83.0, "learning_rate": 0.000195, "loss": 0.0029, "step": 1411 }, { "epoch": 83.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.76246669403774, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6842105263157895, \"recall\": 0.7222222222222222, \"f1-score\": 0.7027027027027027, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8101666927736006, \"recall\": 0.7543949106449106, \"f1-score\": 0.7658868817992679, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7598260698041399, \"recall\": 0.7407407407407407, \"f1-score\": 0.7431303044807601, \"support\": 135.0}}", "eval_f1_macro": 76.58868817992679, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.313030448076, "eval_loss": 1.025146722793579, "eval_runtime": 2.189, "eval_samples_per_second": 61.672, "step": 1411 }, { "epoch": 84.0, "learning_rate": 0.00019333333333333333, "loss": 0.0008, "step": 1428 }, { "epoch": 84.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.65410439323483, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.766517611157791, \"recall\": 0.7387031371406372, \"f1-score\": 0.7312475747258356, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7362255787043392, \"recall\": 0.7111111111111111, \"f1-score\": 0.7126943787813353, \"support\": 135.0}}", "eval_f1_macro": 73.12475747258355, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.26943787813353, "eval_loss": 1.0640238523483276, "eval_runtime": 2.1727, "eval_samples_per_second": 62.136, "step": 1428 }, { "epoch": 85.0, "learning_rate": 0.00019166666666666665, "loss": 0.0007, "step": 1445 }, { "epoch": 85.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.63498328794381, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7612767620310724, \"recall\": 0.7369548853923854, \"f1-score\": 0.730588001969581, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7325891775317063, \"recall\": 0.7111111111111111, \"f1-score\": 0.7125891073259495, \"support\": 135.0}}", "eval_f1_macro": 73.05880019695809, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.25891073259494, "eval_loss": 1.0694376230239868, "eval_runtime": 2.1511, "eval_samples_per_second": 62.759, "step": 1445 }, { "epoch": 86.0, "learning_rate": 0.00018999999999999998, "loss": 0.0013, "step": 1462 }, { "epoch": 86.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.88851983050755, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6944444444444444, \"recall\": 0.6944444444444444, \"f1-score\": 0.6944444444444444, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7967806461352657, \"recall\": 0.7448462995337995, \"f1-score\": 0.7532179628803489, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.754411567364466, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356561636732861, \"support\": 135.0}}", "eval_f1_macro": 75.3217962880349, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.56561636732862, "eval_loss": 1.076446771621704, "eval_runtime": 2.2436, "eval_samples_per_second": 60.171, "step": 1462 }, { "epoch": 87.0, "learning_rate": 0.00018833333333333332, "loss": 0.0052, "step": 1479 }, { "epoch": 87.0, "eval_accuracy": 77.03703703703704, "eval_average_metrics": 77.53590008051296, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7941176470588235, \"recall\": 0.6923076923076923, \"f1-score\": 0.7397260273972601, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7703703703703704, \"macro avg\": {\"precision\": 0.833837516113058, \"recall\": 0.7720898892773893, \"f1-score\": 0.7884724218577965, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.784561747926041, \"recall\": 0.7703703703703704, \"f1-score\": 0.7722228406219811, \"support\": 135.0}}", "eval_f1_macro": 78.84724218577965, "eval_f1_micro": 77.03703703703704, "eval_f1_weighted": 77.22228406219811, "eval_loss": 1.0223504304885864, "eval_runtime": 2.1607, "eval_samples_per_second": 62.479, "step": 1479 }, { "epoch": 88.0, "learning_rate": 0.00018666666666666666, "loss": 0.0004, "step": 1496 }, { "epoch": 88.0, "eval_accuracy": 75.55555555555556, "eval_average_metrics": 76.05592523834689, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6923076923076923, \"f1-score\": 0.7199999999999999, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8130099067599068, \"recall\": 0.7651454448329449, \"f1-score\": 0.773336418091264, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7716006216006216, \"recall\": 0.7555555555555555, \"f1-score\": 0.7577894803315004, \"support\": 135.0}}", "eval_f1_macro": 77.33364180912639, "eval_f1_micro": 75.55555555555556, "eval_f1_weighted": 75.77894803315003, "eval_loss": 1.0519832372665405, "eval_runtime": 2.2144, "eval_samples_per_second": 60.966, "step": 1496 }, { "epoch": 89.0, "learning_rate": 0.000185, "loss": 0.0004, "step": 1513 }, { "epoch": 89.0, "eval_accuracy": 75.55555555555556, "eval_average_metrics": 76.11648469654651, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7714285714285715, \"recall\": 0.6923076923076923, \"f1-score\": 0.7297297297297296, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8263431013431014, \"recall\": 0.7546620046620047, \"f1-score\": 0.7756144827692142, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7708913308913309, \"recall\": 0.7555555555555555, \"f1-score\": 0.7579337939815347, \"support\": 135.0}}", "eval_f1_macro": 77.56144827692142, "eval_f1_micro": 75.55555555555556, "eval_f1_weighted": 75.79337939815348, "eval_loss": 1.0824300050735474, "eval_runtime": 2.2212, "eval_samples_per_second": 60.777, "step": 1513 }, { "epoch": 90.0, "learning_rate": 0.00018333333333333334, "loss": 0.0023, "step": 1530 }, { "epoch": 90.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89906381584439, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7837301587301587, \"recall\": 0.7384360431235432, \"f1-score\": 0.7390882467117281, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7427689594356262, \"recall\": 0.7111111111111111, \"f1-score\": 0.7146520836998245, \"support\": 135.0}}", "eval_f1_macro": 73.90882467117281, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.46520836998246, "eval_loss": 1.1680539846420288, "eval_runtime": 2.2118, "eval_samples_per_second": 61.037, "step": 1530 }, { "epoch": 91.0, "learning_rate": 0.00018166666666666665, "loss": 0.0017, "step": 1547 }, { "epoch": 91.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.54780705462385, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7821969696969697, \"recall\": 0.8073462995337995, \"f1-score\": 0.7829353295805501, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7491021324354658, \"recall\": 0.7185185185185186, \"f1-score\": 0.7219399155673665, \"support\": 135.0}}", "eval_f1_macro": 78.29353295805501, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.19399155673665, "eval_loss": 1.1076750755310059, "eval_runtime": 2.1138, "eval_samples_per_second": 63.867, "step": 1547 }, { "epoch": 92.0, "learning_rate": 0.00017999999999999998, "loss": 0.0072, "step": 1564 }, { "epoch": 92.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.88573213660933, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7721699167819858, \"recall\": 0.7942344114219114, \"f1-score\": 0.7758369725875447, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.742619934459015, \"recall\": 0.725925925925926, \"f1-score\": 0.7277404610249767, \"support\": 135.0}}", "eval_f1_macro": 77.58369725875447, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.77404610249766, "eval_loss": 1.0486587285995483, "eval_runtime": 2.2394, "eval_samples_per_second": 60.284, "step": 1564 }, { "epoch": 93.0, "learning_rate": 0.00017833333333333332, "loss": 0.0057, "step": 1581 }, { "epoch": 93.0, "eval_accuracy": 68.88888888888889, "eval_average_metrics": 69.75310065197965, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.5, \"f1-score\": 0.5806451612903226, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5, \"recall\": 0.7777777777777778, \"f1-score\": 0.6086956521739131, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6888888888888889, \"macro avg\": {\"precision\": 0.7741794389820706, \"recall\": 0.7117023601398601, \"f1-score\": 0.7183544685246277, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7288036330141594, \"recall\": 0.6888888888888889, \"f1-score\": 0.693991779776781, \"support\": 135.0}}", "eval_f1_macro": 71.83544685246277, "eval_f1_micro": 68.88888888888889, "eval_f1_weighted": 69.3991779776781, "eval_loss": 1.1878422498703003, "eval_runtime": 2.1986, "eval_samples_per_second": 61.403, "step": 1581 }, { "epoch": 94.0, "learning_rate": 0.00017666666666666666, "loss": 0.0195, "step": 1598 }, { "epoch": 94.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.61304930742271, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7979249011857708, \"recall\": 0.7587351884226884, \"f1-score\": 0.7599361706778587, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7623871565900551, \"recall\": 0.7407407407407407, \"f1-score\": 0.7431043201375682, \"support\": 135.0}}", "eval_f1_macro": 75.99361706778586, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.31043201375682, "eval_loss": 1.0754647254943848, "eval_runtime": 2.2267, "eval_samples_per_second": 60.629, "step": 1598 }, { "epoch": 95.0, "learning_rate": 0.000175, "loss": 0.0009, "step": 1615 }, { "epoch": 95.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.42176515283398, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.7272727272727273, \"f1-score\": 0.6956521739130435, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7921437115516063, \"recall\": 0.7429523601398601, \"f1-score\": 0.7544827980154067, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7502160282862037, \"recall\": 0.7407407407407407, \"f1-score\": 0.7409063266164715, \"support\": 135.0}}", "eval_f1_macro": 75.44827980154068, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.09063266164715, "eval_loss": 1.0278376340866089, "eval_runtime": 2.155, "eval_samples_per_second": 62.646, "step": 1615 }, { "epoch": 96.0, "learning_rate": 0.0001733333333333333, "loss": 0.0004, "step": 1632 }, { "epoch": 96.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 75.17056142900891, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7297297297297297, \"recall\": 0.75, \"f1-score\": 0.7397260273972601, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7977492664992665, \"recall\": 0.7543159965034965, \"f1-score\": 0.7625258134847175, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7574262074262075, \"recall\": 0.7481481481481481, \"f1-score\": 0.7480003473793427, \"support\": 135.0}}", "eval_f1_macro": 76.25258134847175, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 74.80003473793427, "eval_loss": 1.0315518379211426, "eval_runtime": 2.137, "eval_samples_per_second": 63.172, "step": 1632 }, { "epoch": 97.0, "learning_rate": 0.00017166666666666665, "loss": 0.0004, "step": 1649 }, { "epoch": 97.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 75.17542277410698, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7647058823529411, \"recall\": 0.7222222222222222, \"f1-score\": 0.7428571428571428, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7985009658384271, \"recall\": 0.7494415306915307, \"f1-score\": 0.7611168071694387, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7574858211390718, \"recall\": 0.7481481481481481, \"f1-score\": 0.7496038074985443, \"support\": 135.0}}", "eval_f1_macro": 76.11168071694388, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 74.96038074985442, "eval_loss": 1.0185447931289673, "eval_runtime": 2.2095, "eval_samples_per_second": 61.099, "step": 1649 }, { "epoch": 98.0, "learning_rate": 0.00016999999999999999, "loss": 0.0081, "step": 1666 }, { "epoch": 98.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.67129308878316, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7769345152528264, \"recall\": 0.8063993298368298, \"f1-score\": 0.784722770729386, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7505437952307024, \"recall\": 0.7333333333333333, \"f1-score\": 0.735462286155274, \"support\": 135.0}}", "eval_f1_macro": 78.4722770729386, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.5462286155274, "eval_loss": 1.0540401935577393, "eval_runtime": 2.1385, "eval_samples_per_second": 63.127, "step": 1666 }, { "epoch": 99.0, "learning_rate": 0.00016833333333333332, "loss": 0.0003, "step": 1683 }, { "epoch": 99.0, "eval_accuracy": 75.55555555555556, "eval_average_metrics": 76.05696734786595, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7428571428571429, \"recall\": 0.7222222222222222, \"f1-score\": 0.732394366197183, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8140947288006112, \"recall\": 0.7654125388500388, \"f1-score\": 0.7737856961026234, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7716894434541492, \"recall\": 0.7555555555555555, \"f1-score\": 0.7573818867009036, \"support\": 135.0}}", "eval_f1_macro": 77.37856961026233, "eval_f1_micro": 75.55555555555556, "eval_f1_weighted": 75.73818867009035, "eval_loss": 1.054269790649414, "eval_runtime": 2.1572, "eval_samples_per_second": 62.58, "step": 1683 }, { "epoch": 100.0, "learning_rate": 0.00016666666666666666, "loss": 0.0016, "step": 1700 }, { "epoch": 100.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 75.19184257143311, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7741935483870968, \"recall\": 0.6666666666666666, \"f1-score\": 0.7164179104477612, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7990435981204232, \"recall\": 0.7535147144522145, \"f1-score\": 0.7606214838883731, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7652767059388452, \"recall\": 0.7481481481481481, \"f1-score\": 0.7507559226726549, \"support\": 135.0}}", "eval_f1_macro": 76.06214838883731, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 75.07559226726548, "eval_loss": 1.0762172937393188, "eval_runtime": 2.2054, "eval_samples_per_second": 61.213, "step": 1700 }, { "epoch": 101.0, "learning_rate": 0.000165, "loss": 0.0008, "step": 1717 }, { "epoch": 101.0, "eval_accuracy": 76.29629629629629, "eval_average_metrics": 76.48385951719736, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8235294117647058, \"recall\": 0.717948717948718, \"f1-score\": 0.767123287671233, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.762962962962963, \"macro avg\": {\"precision\": 0.8140522875816993, \"recall\": 0.748227466977467, \"f1-score\": 0.7677001436238999, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7779084967320261, \"recall\": 0.762962962962963, \"f1-score\": 0.7657283111380684, \"support\": 135.0}}", "eval_f1_macro": 76.77001436238999, "eval_f1_micro": 76.29629629629629, "eval_f1_weighted": 76.57283111380684, "eval_loss": 1.0158424377441406, "eval_runtime": 2.1602, "eval_samples_per_second": 62.494, "step": 1717 }, { "epoch": 102.0, "learning_rate": 0.0001633333333333333, "loss": 0.0005, "step": 1734 }, { "epoch": 102.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 72.98289629063926, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8, \"recall\": 0.7272727272727273, \"f1-score\": 0.761904761904762, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7834636668306023, \"recall\": 0.7290634712509713, \"f1-score\": 0.7370998050821901, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.752118603301399, \"recall\": 0.725925925925926, \"f1-score\": 0.7303641946915284, \"support\": 135.0}}", "eval_f1_macro": 73.709980508219, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 73.03641946915283, "eval_loss": 1.0503590106964111, "eval_runtime": 2.2165, "eval_samples_per_second": 60.907, "step": 1734 }, { "epoch": 103.0, "learning_rate": 0.00016166666666666665, "loss": 0.0013, "step": 1751 }, { "epoch": 103.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.67005589980079, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8069940476190476, \"recall\": 0.752324932012432, \"f1-score\": 0.7613581176647959, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7638364975619877, \"recall\": 0.7407407407407407, \"f1-score\": 0.7439626368457539, \"support\": 135.0}}", "eval_f1_macro": 76.13581176647959, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.39626368457539, "eval_loss": 1.0846141576766968, "eval_runtime": 2.2097, "eval_samples_per_second": 61.093, "step": 1751 }, { "epoch": 104.0, "learning_rate": 0.00015999999999999999, "loss": 0.0017, "step": 1768 }, { "epoch": 104.0, "eval_accuracy": 67.4074074074074, "eval_average_metrics": 67.85678034161822, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.7272727272727273, \"recall\": 0.8888888888888888, \"f1-score\": 0.7999999999999999, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6428571428571429, \"recall\": 0.8181818181818182, \"f1-score\": 0.7200000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.65, \"recall\": 0.6666666666666666, \"f1-score\": 0.6582278481012659, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.6153846153846154, \"f1-score\": 0.761904761904762, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.6875, \"f1-score\": 0.5789473684210527, \"support\": 16.0}, \"accuracy\": 0.674074074074074, \"macro avg\": {\"precision\": 0.7351551226551227, \"recall\": 0.6922166375291375, \"f1-score\": 0.689884997303385, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7053102453102452, \"recall\": 0.674074074074074, \"f1-score\": 0.6762380682131959, \"support\": 135.0}}", "eval_f1_macro": 68.9884997303385, "eval_f1_micro": 67.4074074074074, "eval_f1_weighted": 67.62380682131959, "eval_loss": 1.1709874868392944, "eval_runtime": 2.145, "eval_samples_per_second": 62.936, "step": 1768 }, { "epoch": 105.0, "learning_rate": 0.00015833333333333332, "loss": 0.0039, "step": 1785 }, { "epoch": 105.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.57626910983969, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6410256410256411, \"recall\": 0.6410256410256411, \"f1-score\": 0.6410256410256411, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.6875, \"f1-score\": 0.5789473684210527, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7753345792138895, \"recall\": 0.7230781371406372, \"f1-score\": 0.7304419012808433, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7212663028754984, \"recall\": 0.6962962962962963, \"f1-score\": 0.7000162705201517, \"support\": 135.0}}", "eval_f1_macro": 73.04419012808432, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 70.00162705201517, "eval_loss": 1.1890654563903809, "eval_runtime": 2.2179, "eval_samples_per_second": 60.869, "step": 1785 }, { "epoch": 106.0, "learning_rate": 0.00015666666666666666, "loss": 0.0005, "step": 1802 }, { "epoch": 106.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.60340419650365, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7788739038739039, \"recall\": 0.7317586926961928, \"f1-score\": 0.7311720214719142, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7352949519616188, \"recall\": 0.6962962962962963, \"f1-score\": 0.7003715537956388, \"support\": 135.0}}", "eval_f1_macro": 73.11720214719142, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 70.03715537956387, "eval_loss": 1.2529244422912598, "eval_runtime": 2.2315, "eval_samples_per_second": 60.498, "step": 1802 }, { "epoch": 107.0, "learning_rate": 0.000155, "loss": 0.001, "step": 1819 }, { "epoch": 107.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.58287100496284, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.5555555555555556, \"f1-score\": 0.6250000000000001, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7766354016354016, \"recall\": 0.7317586926961928, \"f1-score\": 0.7314957452929186, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7284128765610247, \"recall\": 0.6962962962962963, \"f1-score\": 0.6992265023130029, \"support\": 135.0}}", "eval_f1_macro": 73.14957452929185, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.92265023130028, "eval_loss": 1.2114461660385132, "eval_runtime": 2.1003, "eval_samples_per_second": 64.278, "step": 1819 }, { "epoch": 108.0, "learning_rate": 0.0001533333333333333, "loss": 0.0002, "step": 1836 }, { "epoch": 108.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.58367232228676, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.8125, \"f1-score\": 0.6190476190476191, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.77743083993084, \"recall\": 0.7317586926961928, \"f1-score\": 0.7323127643359169, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7265083265083264, \"recall\": 0.6962962962962963, \"f1-score\": 0.698441535962961, \"support\": 135.0}}", "eval_f1_macro": 73.23127643359169, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.8441535962961, "eval_loss": 1.21000075340271, "eval_runtime": 2.1341, "eval_samples_per_second": 63.257, "step": 1836 }, { "epoch": 109.0, "learning_rate": 0.00015166666666666665, "loss": 0.0041, "step": 1853 }, { "epoch": 109.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.92307044896559, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7983946608946609, \"recall\": 0.744779526029526, \"f1-score\": 0.7552466149013464, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7484559884559885, \"recall\": 0.7333333333333333, \"f1-score\": 0.7350095363906106, \"support\": 135.0}}", "eval_f1_macro": 75.52466149013463, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.50095363906107, "eval_loss": 1.1299362182617188, "eval_runtime": 2.2527, "eval_samples_per_second": 59.928, "step": 1853 }, { "epoch": 110.0, "learning_rate": 0.00015, "loss": 0.0012, "step": 1870 }, { "epoch": 110.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.93367080496068, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.798580771374889, \"recall\": 0.744512432012432, \"f1-score\": 0.7553214077392971, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7485956238897415, \"recall\": 0.7333333333333333, \"f1-score\": 0.7353587577924635, \"support\": 135.0}}", "eval_f1_macro": 75.53214077392971, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.53587577924635, "eval_loss": 1.1259392499923706, "eval_runtime": 2.1641, "eval_samples_per_second": 62.382, "step": 1870 }, { "epoch": 111.0, "learning_rate": 0.00014833333333333332, "loss": 0.0016, "step": 1887 }, { "epoch": 111.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.5803126086232, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7890165755297334, \"recall\": 0.7332277097902098, \"f1-score\": 0.7449370393743795, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7351884762411078, \"recall\": 0.7185185185185186, \"f1-score\": 0.7212384279335117, \"support\": 135.0}}", "eval_f1_macro": 74.49370393743794, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.12384279335117, "eval_loss": 1.1060606241226196, "eval_runtime": 2.1888, "eval_samples_per_second": 61.678, "step": 1887 }, { "epoch": 112.0, "learning_rate": 0.00014666666666666664, "loss": 0.0008, "step": 1904 }, { "epoch": 112.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.58815358410081, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6578947368421053, \"recall\": 0.6410256410256411, \"f1-score\": 0.6493506493506495, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.75, \"f1-score\": 0.6, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.776365629984051, \"recall\": 0.727418414918415, \"f1-score\": 0.7315299716534531, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7238407466477642, \"recall\": 0.6962962962962963, \"f1-score\": 0.6994035791179866, \"support\": 135.0}}", "eval_f1_macro": 73.15299716534531, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.94035791179867, "eval_loss": 1.204077959060669, "eval_runtime": 2.1515, "eval_samples_per_second": 62.746, "step": 1904 }, { "epoch": 113.0, "learning_rate": 0.000145, "loss": 0.0009, "step": 1921 }, { "epoch": 113.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.56981914255991, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7873426359180036, \"recall\": 0.7421753593628594, \"f1-score\": 0.7447551523701981, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7428055060408002, \"recall\": 0.7185185185185186, \"f1-score\": 0.721000576295161, \"support\": 135.0}}", "eval_f1_macro": 74.4755152370198, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.1000576295161, "eval_loss": 1.2206599712371826, "eval_runtime": 2.1932, "eval_samples_per_second": 61.553, "step": 1921 }, { "epoch": 114.0, "learning_rate": 0.00014333333333333334, "loss": 0.0008, "step": 1938 }, { "epoch": 114.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.50895922645417, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7352941176470589, \"recall\": 0.6410256410256411, \"f1-score\": 0.6849315068493151, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7836120335230867, \"recall\": 0.7421753593628594, \"f1-score\": 0.7418505576965078, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7454529482043714, \"recall\": 0.7185185185185186, \"f1-score\": 0.7214707743246217, \"support\": 135.0}}", "eval_f1_macro": 74.18505576965077, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.14707743246217, "eval_loss": 1.201111078262329, "eval_runtime": 2.1314, "eval_samples_per_second": 63.338, "step": 1938 }, { "epoch": 115.0, "learning_rate": 0.00014166666666666665, "loss": 0.0003, "step": 1955 }, { "epoch": 115.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89831172580645, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7827099116161615, \"recall\": 0.7343628593628594, \"f1-score\": 0.7401694157684197, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7314534231200898, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135408310416161, \"support\": 135.0}}", "eval_f1_macro": 74.01694157684197, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.35408310416163, "eval_loss": 1.181020736694336, "eval_runtime": 2.237, "eval_samples_per_second": 60.349, "step": 1955 }, { "epoch": 116.0, "learning_rate": 0.00014, "loss": 0.0002, "step": 1972 }, { "epoch": 116.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.25460880133727, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6578947368421053, \"recall\": 0.6410256410256411, \"f1-score\": 0.6493506493506495, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7801606074952954, \"recall\": 0.7308906371406372, \"f1-score\": 0.7360371831919146, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7287160867554882, \"recall\": 0.7037037037037037, \"f1-score\": 0.7067397614541691, \"support\": 135.0}}", "eval_f1_macro": 73.60371831919146, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.6739761454169, "eval_loss": 1.216983675956726, "eval_runtime": 2.1268, "eval_samples_per_second": 63.474, "step": 1972 }, { "epoch": 117.0, "learning_rate": 0.00013833333333333333, "loss": 0.0015, "step": 1989 }, { "epoch": 117.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.36956241306592, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8001194760365956, \"recall\": 0.7453804875679876, \"f1-score\": 0.7551997801408894, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7443039001734654, \"recall\": 0.725925925925926, \"f1-score\": 0.7277308645298954, \"support\": 135.0}}", "eval_f1_macro": 75.51997801408893, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.77308645298953, "eval_loss": 1.1754764318466187, "eval_runtime": 2.3156, "eval_samples_per_second": 58.299, "step": 1989 }, { "epoch": 118.0, "learning_rate": 0.00013666666666666666, "loss": 0.0011, "step": 2006 }, { "epoch": 118.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.17485160829882, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.5833333333333334, \"f1-score\": 0.6562499999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7760871136872282, \"recall\": 0.7306235431235432, \"f1-score\": 0.7325914102148916, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7313728393819928, \"recall\": 0.7037037037037037, \"f1-score\": 0.7069952467096542, \"support\": 135.0}}", "eval_f1_macro": 73.25914102148916, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.69952467096542, "eval_loss": 1.1750222444534302, "eval_runtime": 2.2858, "eval_samples_per_second": 59.061, "step": 2006 }, { "epoch": 119.0, "learning_rate": 0.000135, "loss": 0.0001, "step": 2023 }, { "epoch": 119.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.73002733490209, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7730584332146831, \"recall\": 0.7262043512043512, \"f1-score\": 0.7334226043582956, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7285210035210035, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135562668155659, \"support\": 135.0}}", "eval_f1_macro": 73.34226043582956, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.35562668155659, "eval_loss": 1.1494803428649902, "eval_runtime": 2.2653, "eval_samples_per_second": 59.593, "step": 2023 }, { "epoch": 120.0, "learning_rate": 0.0001333333333333333, "loss": 0.0004, "step": 2040 }, { "epoch": 120.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.61782293955142, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8034016636957814, \"recall\": 0.752324932012432, \"f1-score\": 0.7609662631209947, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564920917862095, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422651729795805, \"support\": 135.0}}", "eval_f1_macro": 76.09662631209946, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.22651729795805, "eval_loss": 1.1421889066696167, "eval_runtime": 2.1875, "eval_samples_per_second": 61.713, "step": 2040 }, { "epoch": 121.0, "learning_rate": 0.00013166666666666665, "loss": 0.0023, "step": 2057 }, { "epoch": 121.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.5217489806628, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7666666666666667, \"recall\": 0.5897435897435898, \"f1-score\": 0.6666666666666667, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7989348370927318, \"recall\": 0.8121539918414918, \"f1-score\": 0.7989948539871141, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7548801296169717, \"recall\": 0.7407407407407407, \"f1-score\": 0.7403936237579168, \"support\": 135.0}}", "eval_f1_macro": 79.8994853987114, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.03936237579168, "eval_loss": 1.1364786624908447, "eval_runtime": 2.1228, "eval_samples_per_second": 63.596, "step": 2057 }, { "epoch": 122.0, "learning_rate": 0.00013, "loss": 0.0011, "step": 2074 }, { "epoch": 122.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 74.26085371819231, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9090909090909091, \"recall\": 0.7692307692307693, \"f1-score\": 0.8333333333333333, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.875, \"f1-score\": 0.6363636363636364, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7922449945887446, \"recall\": 0.8090156371406372, \"f1-score\": 0.7898259455409633, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7544005531505531, \"recall\": 0.725925925925926, \"f1-score\": 0.728756351334877, \"support\": 135.0}}", "eval_f1_macro": 78.98259455409632, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.87563513348772, "eval_loss": 1.221017599105835, "eval_runtime": 2.2505, "eval_samples_per_second": 59.985, "step": 2074 }, { "epoch": 123.0, "learning_rate": 0.00012833333333333333, "loss": 0.0008, "step": 2091 }, { "epoch": 123.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.57727733650582, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5, \"recall\": 0.875, \"f1-score\": 0.6363636363636364, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7952244116037219, \"recall\": 0.7358318764568764, \"f1-score\": 0.7438368834916149, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7506233655658944, \"recall\": 0.7185185185185186, \"f1-score\": 0.7222171729315804, \"support\": 135.0}}", "eval_f1_macro": 74.38368834916149, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.22171729315805, "eval_loss": 1.2404636144638062, "eval_runtime": 2.218, "eval_samples_per_second": 60.867, "step": 2091 }, { "epoch": 124.0, "learning_rate": 0.00012666666666666666, "loss": 0.0048, "step": 2108 }, { "epoch": 124.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.91986381095504, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7971842356013552, \"recall\": 0.8078804875679876, \"f1-score\": 0.7945727056980254, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7516616475312128, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355551800735092, \"support\": 135.0}}", "eval_f1_macro": 79.45727056980255, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.55551800735093, "eval_loss": 1.1756157875061035, "eval_runtime": 2.1659, "eval_samples_per_second": 62.329, "step": 2108 }, { "epoch": 125.0, "learning_rate": 0.000125, "loss": 0.0007, "step": 2125 }, { "epoch": 125.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.36438346817907, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7759566186985541, \"recall\": 0.7887043512043512, \"f1-score\": 0.7763513430179385, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7347037992199282, \"recall\": 0.7185185185185186, \"f1-score\": 0.721186958672187, \"support\": 135.0}}", "eval_f1_macro": 77.63513430179385, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.11869586721869, "eval_loss": 1.212998628616333, "eval_runtime": 2.2937, "eval_samples_per_second": 58.857, "step": 2125 }, { "epoch": 126.0, "learning_rate": 0.0001233333333333333, "loss": 0.0008, "step": 2142 }, { "epoch": 126.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 71.27406229672498, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8, \"recall\": 0.8888888888888888, \"f1-score\": 0.8421052631578948, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6896551724137931, \"recall\": 0.5555555555555556, \"f1-score\": 0.6153846153846154, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7571729675177952, \"recall\": 0.7828950563325563, \"f1-score\": 0.7595188378961195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7231066799904603, \"recall\": 0.6962962962962963, \"f1-score\": 0.6988510613802864, \"support\": 135.0}}", "eval_f1_macro": 75.95188378961197, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.88510613802865, "eval_loss": 1.2854887247085571, "eval_runtime": 2.074, "eval_samples_per_second": 65.093, "step": 2142 }, { "epoch": 127.0, "learning_rate": 0.00012166666666666665, "loss": 0.0004, "step": 2159 }, { "epoch": 127.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.4292424329562, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.4642857142857143, \"recall\": 0.8125, \"f1-score\": 0.5909090909090908, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7681011118511119, \"recall\": 0.7236001845376845, \"f1-score\": 0.72428547901374, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7304041637374971, \"recall\": 0.6962962962962963, \"f1-score\": 0.7002916257119156, \"support\": 135.0}}", "eval_f1_macro": 72.42854790137399, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 70.02916257119156, "eval_loss": 1.351884365081787, "eval_runtime": 2.1241, "eval_samples_per_second": 63.557, "step": 2159 }, { "epoch": 128.0, "learning_rate": 0.00011999999999999999, "loss": 0.0002, "step": 2176 }, { "epoch": 128.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.41850794610482, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7407407407407407, \"recall\": 0.5555555555555556, \"f1-score\": 0.634920634920635, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7678106428106428, \"recall\": 0.7236001845376845, \"f1-score\": 0.7248628871217477, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7268032082846898, \"recall\": 0.6962962962962963, \"f1-score\": 0.6992848381298528, \"support\": 135.0}}", "eval_f1_macro": 72.48628871217477, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.92848381298528, "eval_loss": 1.3572306632995605, "eval_runtime": 2.1876, "eval_samples_per_second": 61.71, "step": 2176 }, { "epoch": 129.0, "learning_rate": 0.00011833333333333331, "loss": 0.001, "step": 2193 }, { "epoch": 129.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.61782293955142, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8034016636957814, \"recall\": 0.752324932012432, \"f1-score\": 0.7609662631209947, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564920917862095, \"recall\": 0.7407407407407407, \"f1-score\": 0.7422651729795805, \"support\": 135.0}}", "eval_f1_macro": 76.09662631209946, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.22651729795805, "eval_loss": 1.2295937538146973, "eval_runtime": 2.2337, "eval_samples_per_second": 60.438, "step": 2193 }, { "epoch": 130.0, "learning_rate": 0.00011666666666666665, "loss": 0.0028, "step": 2210 }, { "epoch": 130.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.55886340237367, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.75, \"recall\": 0.6153846153846154, \"f1-score\": 0.676056338028169, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7957131716322893, \"recall\": 0.8180300602175602, \"f1-score\": 0.7989535791265285, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7572416508691019, \"recall\": 0.7407407407407407, \"f1-score\": 0.7419194754869369, \"support\": 135.0}}", "eval_f1_macro": 79.89535791265286, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.1919475486937, "eval_loss": 1.259170651435852, "eval_runtime": 2.1066, "eval_samples_per_second": 64.083, "step": 2210 }, { "epoch": 131.0, "learning_rate": 0.000115, "loss": 0.0033, "step": 2227 }, { "epoch": 131.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 72.25202775378126, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6944444444444444, \"recall\": 0.6410256410256411, \"f1-score\": 0.6666666666666666, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7837695868945869, \"recall\": 0.7942586926961928, \"f1-score\": 0.7756004567204229, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7381803313284795, \"recall\": 0.7037037037037037, \"f1-score\": 0.7070732460234199, \"support\": 135.0}}", "eval_f1_macro": 77.56004567204229, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.707324602342, "eval_loss": 1.3266923427581787, "eval_runtime": 2.1701, "eval_samples_per_second": 62.208, "step": 2227 }, { "epoch": 132.0, "learning_rate": 0.00011333333333333331, "loss": 0.0004, "step": 2244 }, { "epoch": 132.0, "eval_accuracy": 69.62962962962963, "eval_average_metrics": 70.6257688194083, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.625, \"recall\": 0.9090909090909091, \"f1-score\": 0.7407407407407406, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6756756756756757, \"recall\": 0.6410256410256411, \"f1-score\": 0.6578947368421052, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.6962962962962963, \"macro avg\": {\"precision\": 0.7814234907984907, \"recall\": 0.7317586926961928, \"f1-score\": 0.7328372988256859, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327582425730573, \"recall\": 0.6962962962962963, \"f1-score\": 0.699600861358053, \"support\": 135.0}}", "eval_f1_macro": 73.28372988256861, "eval_f1_micro": 69.62962962962963, "eval_f1_weighted": 69.9600861358053, "eval_loss": 1.33100426197052, "eval_runtime": 2.1896, "eval_samples_per_second": 61.654, "step": 2244 }, { "epoch": 133.0, "learning_rate": 0.00011166666666666667, "loss": 0.0006, "step": 2261 }, { "epoch": 133.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.70429974221537, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7960415628762403, \"recall\": 0.7419082653457654, \"f1-score\": 0.7506471755169501, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7390933483406602, \"recall\": 0.7185185185185186, \"f1-score\": 0.7204877771346277, \"support\": 135.0}}", "eval_f1_macro": 75.06471755169501, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.04877771346277, "eval_loss": 1.2336255311965942, "eval_runtime": 2.2249, "eval_samples_per_second": 60.678, "step": 2261 }, { "epoch": 134.0, "learning_rate": 0.00010999999999999998, "loss": 0.0001, "step": 2278 }, { "epoch": 134.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 72.02848654875856, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7692307692307693, \"recall\": 0.5555555555555556, \"f1-score\": 0.6451612903225806, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.48148148148148145, \"recall\": 0.8125, \"f1-score\": 0.6046511627906976, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7946632071632072, \"recall\": 0.7381689491064491, \"f1-score\": 0.7445463256662919, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7448381796529944, \"recall\": 0.7111111111111111, \"f1-score\": 0.7143709140618288, \"support\": 135.0}}", "eval_f1_macro": 74.45463256662919, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.43709140618289, "eval_loss": 1.3270906209945679, "eval_runtime": 2.1891, "eval_samples_per_second": 61.669, "step": 2278 }, { "epoch": 135.0, "learning_rate": 0.00010833333333333333, "loss": 0.0014, "step": 2295 }, { "epoch": 135.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.72547585668991, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6944444444444444, \"f1-score\": 0.7042253521126761, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8124975198412698, \"recall\": 0.7569323038073038, \"f1-score\": 0.7644743043935498, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7664916225749558, \"recall\": 0.7407407407407407, \"f1-score\": 0.7430632483925649, \"support\": 135.0}}", "eval_f1_macro": 76.44743043935497, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.30632483925649, "eval_loss": 1.251107931137085, "eval_runtime": 2.1766, "eval_samples_per_second": 62.023, "step": 2295 }, { "epoch": 136.0, "learning_rate": 0.00010666666666666667, "loss": 0.0024, "step": 2312 }, { "epoch": 136.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 76.36714659524762, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6842105263157895, \"recall\": 0.7222222222222222, \"f1-score\": 0.7027027027027027, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8275862068965517, \"recall\": 0.6153846153846154, \"f1-score\": 0.7058823529411765, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.8174468138737649, \"recall\": 0.8196993978243978, \"f1-score\": 0.8082919113289957, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7746230198740786, \"recall\": 0.7481481481481481, \"f1-score\": 0.7500976561846127, \"support\": 135.0}}", "eval_f1_macro": 80.82919113289957, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 75.00976561846126, "eval_loss": 1.2201787233352661, "eval_runtime": 2.1085, "eval_samples_per_second": 64.026, "step": 2312 }, { "epoch": 137.0, "learning_rate": 0.00010499999999999999, "loss": 0.0006, "step": 2329 }, { "epoch": 137.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 73.34865168139474, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.76, \"recall\": 0.5277777777777778, \"f1-score\": 0.6229508196721312, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5384615384615384, \"recall\": 0.875, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7731202865761689, \"recall\": 0.8114194832944832, \"f1-score\": 0.7785896412055908, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7473708731355789, \"recall\": 0.7185185185185186, \"f1-score\": 0.7183193890131623, \"support\": 135.0}}", "eval_f1_macro": 77.85896412055907, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 71.83193890131622, "eval_loss": 1.2545726299285889, "eval_runtime": 2.0734, "eval_samples_per_second": 65.11, "step": 2329 }, { "epoch": 138.0, "learning_rate": 0.00010333333333333333, "loss": 0.0018, "step": 2346 }, { "epoch": 138.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.43196457326891, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7872411405020101, \"recall\": 0.809604458041958, \"f1-score\": 0.792512077294686, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7577424186119838, \"recall\": 0.7407407407407407, \"f1-score\": 0.7432850241545893, \"support\": 135.0}}", "eval_f1_macro": 79.2512077294686, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.32850241545893, "eval_loss": 1.1618794202804565, "eval_runtime": 2.1519, "eval_samples_per_second": 62.735, "step": 2346 }, { "epoch": 139.0, "learning_rate": 0.00010166666666666667, "loss": 0.0004, "step": 2363 }, { "epoch": 139.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.73814060272393, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7714285714285715, \"recall\": 0.6923076923076923, \"f1-score\": 0.7297297297297296, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7857676140284836, \"recall\": 0.7979737276612278, \"f1-score\": 0.7869864182364182, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7477583286278938, \"recall\": 0.7333333333333333, \"f1-score\": 0.7358725392058726, \"support\": 135.0}}", "eval_f1_macro": 78.69864182364182, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.58725392058726, "eval_loss": 1.1620222330093384, "eval_runtime": 2.2656, "eval_samples_per_second": 59.586, "step": 2363 }, { "epoch": 140.0, "learning_rate": 9.999999999999999e-05, "loss": 0.0044, "step": 2380 }, { "epoch": 140.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.40867003367005, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7272727272727273, \"recall\": 0.7272727272727273, \"f1-score\": 0.7272727272727273, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7, \"recall\": 0.5833333333333334, \"f1-score\": 0.6363636363636365, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7788792530568847, \"recall\": 0.7317344114219114, \"f1-score\": 0.7391414141414141, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7357835666607596, \"recall\": 0.7185185185185186, \"f1-score\": 0.7201683501683502, \"support\": 135.0}}", "eval_f1_macro": 73.91414141414143, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.01683501683502, "eval_loss": 1.2089694738388062, "eval_runtime": 2.2372, "eval_samples_per_second": 60.342, "step": 2380 }, { "epoch": 141.0, "learning_rate": 9.833333333333333e-05, "loss": 0.0005, "step": 2397 }, { "epoch": 141.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.0998434515311, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7567567567567568, \"recall\": 0.717948717948718, \"f1-score\": 0.736842105263158, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.780597947085016, \"recall\": 0.7430980477855478, \"f1-score\": 0.7443116919947583, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7462899681290487, \"recall\": 0.725925925925926, \"f1-score\": 0.7278301942146337, \"support\": 135.0}}", "eval_f1_macro": 74.43116919947583, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.78301942146337, "eval_loss": 1.2010650634765625, "eval_runtime": 2.2981, "eval_samples_per_second": 58.744, "step": 2397 }, { "epoch": 142.0, "learning_rate": 9.666666666666667e-05, "loss": 0.0001, "step": 2414 }, { "epoch": 142.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 74.19280926182283, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7241379310344828, \"recall\": 0.5833333333333334, \"f1-score\": 0.6461538461538462, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7850638100067613, \"recall\": 0.8105514277389277, \"f1-score\": 0.7877778983036839, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7492036661407858, \"recall\": 0.725925925925926, \"f1-score\": 0.7280826203173772, \"support\": 135.0}}", "eval_f1_macro": 78.77778983036839, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.80826203173773, "eval_loss": 1.2088888883590698, "eval_runtime": 2.1385, "eval_samples_per_second": 63.128, "step": 2414 }, { "epoch": 143.0, "learning_rate": 9.499999999999999e-05, "loss": 0.001, "step": 2431 }, { "epoch": 143.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.53753615871489, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.5833333333333334, \"recall\": 0.7777777777777778, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.793443807141625, \"recall\": 0.8174958721833722, \"f1-score\": 0.797105878846315, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7597858193999876, \"recall\": 0.7407407407407407, \"f1-score\": 0.7429140860207989, \"support\": 135.0}}", "eval_f1_macro": 79.7105878846315, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.29140860207988, "eval_loss": 1.1494632959365845, "eval_runtime": 2.0877, "eval_samples_per_second": 64.663, "step": 2431 }, { "epoch": 144.0, "learning_rate": 9.333333333333333e-05, "loss": 0.0003, "step": 2448 }, { "epoch": 144.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.5783020422968, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7142857142857143, \"recall\": 0.6944444444444444, \"f1-score\": 0.7042253521126761, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7575757575757576, \"recall\": 0.6410256410256411, \"f1-score\": 0.6944444444444444, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7968614718614719, \"recall\": 0.8134226884226884, \"f1-score\": 0.7990512962136822, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7549655282988617, \"recall\": 0.7407407407407407, \"f1-score\": 0.742599303996708, \"support\": 135.0}}", "eval_f1_macro": 79.90512962136823, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.25993039967081, "eval_loss": 1.1091960668563843, "eval_runtime": 2.1685, "eval_samples_per_second": 62.254, "step": 2448 }, { "epoch": 145.0, "learning_rate": 9.166666666666667e-05, "loss": 0.0002, "step": 2465 }, { "epoch": 145.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 76.29216423947909, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6923076923076923, \"recall\": 0.75, \"f1-score\": 0.7199999999999999, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6, \"recall\": 0.75, \"f1-score\": 0.6666666666666665, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.8073717948717949, \"recall\": 0.8171620046620047, \"f1-score\": 0.8057793565410674, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7656030389363723, \"recall\": 0.7481481481481481, \"f1-score\": 0.7496109167418001, \"support\": 135.0}}", "eval_f1_macro": 80.57793565410674, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 74.96109167418001, "eval_loss": 1.1161596775054932, "eval_runtime": 2.1314, "eval_samples_per_second": 63.338, "step": 2465 }, { "epoch": 146.0, "learning_rate": 8.999999999999999e-05, "loss": 0.0001, "step": 2482 }, { "epoch": 146.0, "eval_accuracy": 75.55555555555556, "eval_average_metrics": 76.97719093283912, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6666666666666666, \"recall\": 0.9090909090909091, \"f1-score\": 0.7692307692307692, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.8, \"recall\": 0.6153846153846154, \"f1-score\": 0.6956521739130435, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8120300751879699, \"recall\": 0.8249745046620047, \"f1-score\": 0.8115000772617881, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7727188341223429, \"recall\": 0.7555555555555555, \"f1-score\": 0.7564764489406657, \"support\": 135.0}}", "eval_f1_macro": 81.15000772617881, "eval_f1_micro": 75.55555555555556, "eval_f1_weighted": 75.64764489406657, "eval_loss": 1.1192935705184937, "eval_runtime": 2.1466, "eval_samples_per_second": 62.891, "step": 2482 }, { "epoch": 147.0, "learning_rate": 8.833333333333333e-05, "loss": 0.0004, "step": 2499 }, { "epoch": 147.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.58963733628248, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.7987710818593171, \"recall\": 0.809604458041958, \"f1-score\": 0.7987620895510332, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7557962734433322, \"recall\": 0.7407407407407407, \"f1-score\": 0.7433419224187846, \"support\": 135.0}}", "eval_f1_macro": 79.87620895510332, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.33419224187845, "eval_loss": 1.14371657371521, "eval_runtime": 2.2241, "eval_samples_per_second": 60.698, "step": 2499 }, { "epoch": 148.0, "learning_rate": 8.666666666666665e-05, "loss": 0.0006, "step": 2516 }, { "epoch": 148.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.05647965158836, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7777777777777778, \"recall\": 0.5833333333333334, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7368421052631579, \"recall\": 0.717948717948718, \"f1-score\": 0.7272727272727273, \"support\": 39.0}, \"question\": {\"precision\": 0.5384615384615384, \"recall\": 0.7777777777777778, \"f1-score\": 0.6363636363636364, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7792046221322537, \"recall\": 0.7430980477855478, \"f1-score\": 0.7418917435221783, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7518493527265457, \"recall\": 0.725925925925926, \"f1-score\": 0.7285155906895038, \"support\": 135.0}}", "eval_f1_macro": 74.18917435221783, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.85155906895038, "eval_loss": 1.2226479053497314, "eval_runtime": 2.2006, "eval_samples_per_second": 61.347, "step": 2516 }, { "epoch": 149.0, "learning_rate": 8.499999999999999e-05, "loss": 0.0001, "step": 2533 }, { "epoch": 149.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.12022510871195, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7586206896551724, \"recall\": 0.6111111111111112, \"f1-score\": 0.676923076923077, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7297297297297297, \"recall\": 0.6923076923076923, \"f1-score\": 0.7105263157894737, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7809471629514733, \"recall\": 0.7433651418026419, \"f1-score\": 0.7449589518668466, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7465120614545901, \"recall\": 0.725925925925926, \"f1-score\": 0.7279982006297796, \"support\": 135.0}}", "eval_f1_macro": 74.49589518668466, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.79982006297794, "eval_loss": 1.2157655954360962, "eval_runtime": 2.1514, "eval_samples_per_second": 62.749, "step": 2533 }, { "epoch": 150.0, "learning_rate": 8.333333333333333e-05, "loss": 0.0009, "step": 2550 }, { "epoch": 150.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 76.0197940520429, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7222222222222222, \"recall\": 0.7222222222222222, \"f1-score\": 0.7222222222222222, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7942298714357537, \"recall\": 0.8026600135975136, \"f1-score\": 0.7949344892838044, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7581013031993423, \"recall\": 0.7481481481481481, \"f1-score\": 0.7495609765016158, \"support\": 135.0}}", "eval_f1_macro": 79.49344892838043, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 74.95609765016158, "eval_loss": 1.1733694076538086, "eval_runtime": 2.2975, "eval_samples_per_second": 58.759, "step": 2550 }, { "epoch": 151.0, "learning_rate": 8.166666666666665e-05, "loss": 0.0008, "step": 2567 }, { "epoch": 151.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.8567648637555, "eval_classification_report": "{\"acceptance\": {\"precision\": 0.8888888888888888, \"recall\": 0.8888888888888888, \"f1-score\": 0.8888888888888888, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.625, \"recall\": 0.5555555555555556, \"f1-score\": 0.5882352941176471, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7776136152494848, \"recall\": 0.7818266802641802, \"f1-score\": 0.7749303779482808, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7380173905536225, \"recall\": 0.725925925925926, \"f1-score\": 0.7274883647500868, \"support\": 135.0}}", "eval_f1_macro": 77.49303779482808, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.74883647500869, "eval_loss": 1.1846721172332764, "eval_runtime": 2.2205, "eval_samples_per_second": 60.796, "step": 2567 }, { "epoch": 152.0, "learning_rate": 7.999999999999999e-05, "loss": 0.0001, "step": 2584 }, { "epoch": 152.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 75.57891605906845, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.8020374552983248, \"recall\": 0.8070792055167055, \"f1-score\": 0.7986914184004976, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.755954576824142, \"recall\": 0.7407407407407407, \"f1-score\": 0.7429837424807587, \"support\": 135.0}}", "eval_f1_macro": 79.86914184004976, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.29837424807587, "eval_loss": 1.188364863395691, "eval_runtime": 2.1962, "eval_samples_per_second": 61.471, "step": 2584 }, { "epoch": 153.0, "learning_rate": 7.833333333333333e-05, "loss": 0.0001, "step": 2601 }, { "epoch": 153.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.21584842556148, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6744186046511628, \"recall\": 0.7435897435897436, \"f1-score\": 0.7073170731707318, \"support\": 39.0}, \"question\": {\"precision\": 0.6666666666666666, \"recall\": 0.6666666666666666, \"f1-score\": 0.6666666666666666, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.55, \"recall\": 0.6875, \"f1-score\": 0.6111111111111112, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.8045674770965469, \"recall\": 0.7175905691530691, \"f1-score\": 0.7470421497111972, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7428455093571373, \"recall\": 0.725925925925926, \"f1-score\": 0.7297399354594096, \"support\": 135.0}}", "eval_f1_macro": 74.70421497111973, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.97399354594096, "eval_loss": 1.2244882583618164, "eval_runtime": 2.2128, "eval_samples_per_second": 61.008, "step": 2601 }, { "epoch": 154.0, "learning_rate": 7.666666666666666e-05, "loss": 0.0039, "step": 2618 }, { "epoch": 154.0, "eval_accuracy": 75.55555555555556, "eval_average_metrics": 76.95897870855806, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7647058823529411, \"recall\": 0.7222222222222222, \"f1-score\": 0.7428571428571428, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.631578947368421, \"recall\": 0.75, \"f1-score\": 0.6857142857142857, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8151371074745688, \"recall\": 0.8136897824397824, \"f1-score\": 0.8096817403891035, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.766684030337281, \"recall\": 0.7555555555555555, \"f1-score\": 0.7575662968421079, \"support\": 135.0}}", "eval_f1_macro": 80.96817403891035, "eval_f1_micro": 75.55555555555556, "eval_f1_weighted": 75.75662968421078, "eval_loss": 1.2217811346054077, "eval_runtime": 2.2436, "eval_samples_per_second": 60.172, "step": 2618 }, { "epoch": 155.0, "learning_rate": 7.5e-05, "loss": 0.0003, "step": 2635 }, { "epoch": 155.0, "eval_accuracy": 75.55555555555556, "eval_average_metrics": 77.04653690264276, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7575757575757576, \"recall\": 0.6944444444444444, \"f1-score\": 0.7246376811594203, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7428571428571429, \"recall\": 0.6666666666666666, \"f1-score\": 0.7027027027027027, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8132034632034633, \"recall\": 0.8244403166278166, \"f1-score\": 0.8132795190869966, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7693025493025493, \"recall\": 0.7555555555555555, \"f1-score\": 0.7574708459076027, \"support\": 135.0}}", "eval_f1_macro": 81.32795190869966, "eval_f1_micro": 75.55555555555556, "eval_f1_weighted": 75.74708459076027, "eval_loss": 1.2007545232772827, "eval_runtime": 2.1766, "eval_samples_per_second": 62.023, "step": 2635 }, { "epoch": 156.0, "learning_rate": 7.333333333333332e-05, "loss": 0.0002, "step": 2652 }, { "epoch": 156.0, "eval_accuracy": 75.55555555555556, "eval_average_metrics": 77.04477692135627, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7352941176470589, \"recall\": 0.6944444444444444, \"f1-score\": 0.7142857142857144, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7555555555555555, \"macro avg\": {\"precision\": 0.8131493506493507, \"recall\": 0.8244403166278166, \"f1-score\": 0.8131887812803564, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7696726367314601, \"recall\": 0.7555555555555555, \"f1-score\": 0.7574911844627835, \"support\": 135.0}}", "eval_f1_macro": 81.31887812803565, "eval_f1_micro": 75.55555555555556, "eval_f1_weighted": 75.74911844627835, "eval_loss": 1.1942009925842285, "eval_runtime": 2.3622, "eval_samples_per_second": 57.15, "step": 2652 }, { "epoch": 157.0, "learning_rate": 7.166666666666667e-05, "loss": 0.0022, "step": 2669 }, { "epoch": 157.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.09230915854165, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8025137608806963, \"recall\": 0.7549958721833722, \"f1-score\": 0.7614065508371455, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7533344711839336, \"recall\": 0.7333333333333333, \"f1-score\": 0.735619148837854, \"support\": 135.0}}", "eval_f1_macro": 76.14065508371455, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.56191488378542, "eval_loss": 1.217727541923523, "eval_runtime": 2.2066, "eval_samples_per_second": 61.18, "step": 2669 }, { "epoch": 158.0, "learning_rate": 7e-05, "loss": 0.0, "step": 2686 }, { "epoch": 158.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.62519417356738, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", "eval_f1_macro": 74.72856934535828, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06850364520756, "eval_loss": 1.2388432025909424, "eval_runtime": 2.153, "eval_samples_per_second": 62.703, "step": 2686 }, { "epoch": 159.0, "learning_rate": 6.833333333333333e-05, "loss": 0.0006, "step": 2703 }, { "epoch": 159.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.02792297951471, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.75, \"recall\": 0.6666666666666666, \"f1-score\": 0.7058823529411765, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040257260365956, \"recall\": 0.7488527097902098, \"f1-score\": 0.7588762507291247, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7526372335067987, \"recall\": 0.7333333333333333, \"f1-score\": 0.7355740017847974, \"support\": 135.0}}", "eval_f1_macro": 75.88762507291247, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.55740017847972, "eval_loss": 1.221247673034668, "eval_runtime": 2.273, "eval_samples_per_second": 59.392, "step": 2703 }, { "epoch": 160.0, "learning_rate": 6.666666666666666e-05, "loss": 0.0028, "step": 2720 }, { "epoch": 160.0, "eval_accuracy": 74.81481481481481, "eval_average_metrics": 76.18622883029525, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 1.0, \"f1-score\": 1.0, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7428571428571429, \"recall\": 0.7222222222222222, \"f1-score\": 0.732394366197183, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7741935483870968, \"recall\": 0.6153846153846154, \"f1-score\": 0.6857142857142857, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.8461538461538461, \"recall\": 0.8461538461538461, \"f1-score\": 0.8461538461538461, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7481481481481481, \"macro avg\": {\"precision\": 0.7960499178241114, \"recall\": 0.8215022824397824, \"f1-score\": 0.8033114500290621, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7596155099380906, \"recall\": 0.7481481481481481, \"f1-score\": 0.7478414068864512, \"support\": 135.0}}", "eval_f1_macro": 80.33114500290621, "eval_f1_micro": 74.81481481481481, "eval_f1_weighted": 74.78414068864512, "eval_loss": 1.1919903755187988, "eval_runtime": 2.2082, "eval_samples_per_second": 61.136, "step": 2720 }, { "epoch": 161.0, "learning_rate": 6.5e-05, "loss": 0.0005, "step": 2737 }, { "epoch": 161.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.24915577342048, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7860111205219901, \"recall\": 0.75152364996115, \"f1-score\": 0.7518137254901962, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7415646457675443, \"recall\": 0.725925925925926, \"f1-score\": 0.7263006535947714, \"support\": 135.0}}", "eval_f1_macro": 75.1813725490196, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.63006535947713, "eval_loss": 1.2137112617492676, "eval_runtime": 2.2438, "eval_samples_per_second": 60.166, "step": 2737 }, { "epoch": 162.0, "learning_rate": 6.333333333333333e-05, "loss": 0.0003, "step": 2754 }, { "epoch": 162.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 74.01724012020507, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7419354838709677, \"recall\": 0.6388888888888888, \"f1-score\": 0.6865671641791045, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.797503056005861, \"recall\": 0.7549958721833722, \"f1-score\": 0.7588846210125841, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7518832439356048, \"recall\": 0.7333333333333333, \"f1-score\": 0.7351383171289522, \"support\": 135.0}}", "eval_f1_macro": 75.88846210125841, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.51383171289521, "eval_loss": 1.2145878076553345, "eval_runtime": 2.142, "eval_samples_per_second": 63.026, "step": 2754 }, { "epoch": 163.0, "learning_rate": 6.166666666666666e-05, "loss": 0.0007, "step": 2771 }, { "epoch": 163.0, "eval_accuracy": 74.07407407407408, "eval_average_metrics": 74.5749463235295, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7027027027027027, \"recall\": 0.7222222222222222, \"f1-score\": 0.7123287671232876, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 0.9166666666666666, \"recall\": 0.8461538461538461, \"f1-score\": 0.8799999999999999, \"support\": 13.0}, \"request\": {\"precision\": 0.5909090909090909, \"recall\": 0.8125, \"f1-score\": 0.6842105263157896, \"support\": 16.0}, \"accuracy\": 0.7407407407407407, \"macro avg\": {\"precision\": 0.797979519073269, \"recall\": 0.7508437742812742, \"f1-score\": 0.7594312360087268, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.755945511362178, \"recall\": 0.7407407407407407, \"f1-score\": 0.7420851354509721, \"support\": 135.0}}", "eval_f1_macro": 75.94312360087267, "eval_f1_micro": 74.07407407407408, "eval_f1_weighted": 74.2085135450972, "eval_loss": 1.2377903461456299, "eval_runtime": 2.1826, "eval_samples_per_second": 61.852, "step": 2771 }, { "epoch": 164.0, "learning_rate": 5.9999999999999995e-05, "loss": 0.0012, "step": 2788 }, { "epoch": 164.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.9473371413507, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7955596194087242, \"recall\": 0.747104458041958, \"f1-score\": 0.7549024404282261, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7527513312679553, \"recall\": 0.7333333333333333, \"f1-score\": 0.7363243785591355, \"support\": 135.0}}", "eval_f1_macro": 75.49024404282261, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.63243785591355, "eval_loss": 1.2320207357406616, "eval_runtime": 2.1945, "eval_samples_per_second": 61.519, "step": 2788 }, { "epoch": 165.0, "learning_rate": 5.8333333333333326e-05, "loss": 0.0002, "step": 2805 }, { "epoch": 165.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.35054466230937, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7934839466089466, \"recall\": 0.75152364996115, \"f1-score\": 0.7543137254901962, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7467981401314735, \"recall\": 0.725925925925926, \"f1-score\": 0.7278562091503267, \"support\": 135.0}}", "eval_f1_macro": 75.43137254901961, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.78562091503268, "eval_loss": 1.2451300621032715, "eval_runtime": 2.1694, "eval_samples_per_second": 62.23, "step": 2805 }, { "epoch": 166.0, "learning_rate": 5.666666666666666e-05, "loss": 0.0004, "step": 2822 }, { "epoch": 166.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.29745663811057, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7647058823529411, \"recall\": 0.6666666666666666, \"f1-score\": 0.7123287671232877, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.7926096513780337, \"recall\": 0.7479725135975136, \"f1-score\": 0.7521030279803201, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7472310129957188, \"recall\": 0.725925925925926, \"f1-score\": 0.7279433856922508, \"support\": 135.0}}", "eval_f1_macro": 75.21030279803202, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.79433856922508, "eval_loss": 1.2745343446731567, "eval_runtime": 2.2435, "eval_samples_per_second": 60.174, "step": 2822 }, { "epoch": 167.0, "learning_rate": 5.499999999999999e-05, "loss": 0.0001, "step": 2839 }, { "epoch": 167.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.96219025098748, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.696969696969697, \"recall\": 0.6388888888888888, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7878787878787878, \"recall\": 0.6666666666666666, \"f1-score\": 0.7222222222222221, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.7966899766899767, \"recall\": 0.7514447358197358, \"f1-score\": 0.7557906902598437, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7564506604506605, \"recall\": 0.7333333333333333, \"f1-score\": 0.7360302531129886, \"support\": 135.0}}", "eval_f1_macro": 75.57906902598438, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.60302531129886, "eval_loss": 1.2781518697738647, "eval_runtime": 2.2619, "eval_samples_per_second": 59.684, "step": 2839 }, { "epoch": 168.0, "learning_rate": 5.333333333333333e-05, "loss": 0.0002, "step": 2856 }, { "epoch": 168.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.56451583121837, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.56, \"recall\": 0.875, \"f1-score\": 0.6829268292682927, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7900713712810488, \"recall\": 0.738357128982129, \"f1-score\": 0.7450481168947851, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740871934119246, \"recall\": 0.7185185185185186, \"f1-score\": 0.7204954793169126, \"support\": 135.0}}", "eval_f1_macro": 74.50481168947852, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.04954793169127, "eval_loss": 1.305252194404602, "eval_runtime": 2.1319, "eval_samples_per_second": 63.325, "step": 2856 }, { "epoch": 169.0, "learning_rate": 5.1666666666666664e-05, "loss": 0.0001, "step": 2873 }, { "epoch": 169.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89929243227385, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", "eval_f1_macro": 73.97919123590555, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.39575627096765, "eval_loss": 1.3155133724212646, "eval_runtime": 2.2317, "eval_samples_per_second": 60.491, "step": 2873 }, { "epoch": 170.0, "learning_rate": 4.9999999999999996e-05, "loss": 0.0, "step": 2890 }, { "epoch": 170.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89929243227385, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", "eval_f1_macro": 73.97919123590555, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.39575627096765, "eval_loss": 1.3125479221343994, "eval_runtime": 2.1736, "eval_samples_per_second": 62.109, "step": 2890 }, { "epoch": 171.0, "learning_rate": 4.8333333333333334e-05, "loss": 0.0001, "step": 2907 }, { "epoch": 171.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89929243227385, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", "eval_f1_macro": 73.97919123590555, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.39575627096765, "eval_loss": 1.3118623495101929, "eval_runtime": 2.1217, "eval_samples_per_second": 63.628, "step": 2907 }, { "epoch": 172.0, "learning_rate": 4.6666666666666665e-05, "loss": 0.0, "step": 2924 }, { "epoch": 172.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89929243227385, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", "eval_f1_macro": 73.97919123590555, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.39575627096765, "eval_loss": 1.312318205833435, "eval_runtime": 2.0073, "eval_samples_per_second": 67.255, "step": 2924 }, { "epoch": 173.0, "learning_rate": 4.4999999999999996e-05, "loss": 0.0, "step": 2941 }, { "epoch": 173.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89929243227385, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", "eval_f1_macro": 73.97919123590555, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.39575627096765, "eval_loss": 1.3142341375350952, "eval_runtime": 2.155, "eval_samples_per_second": 62.645, "step": 2941 }, { "epoch": 174.0, "learning_rate": 4.333333333333333e-05, "loss": 0.0001, "step": 2958 }, { "epoch": 174.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.9747073753716, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040708376645876, \"recall\": 0.7455686674436675, \"f1-score\": 0.7566759148589195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7558651279484612, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356457134892781, \"support\": 135.0}}", "eval_f1_macro": 75.66759148589195, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.5645713489278, "eval_loss": 1.2926682233810425, "eval_runtime": 2.0992, "eval_samples_per_second": 64.311, "step": 2958 }, { "epoch": 175.0, "learning_rate": 4.1666666666666665e-05, "loss": 0.0001, "step": 2975 }, { "epoch": 175.0, "eval_accuracy": 73.33333333333333, "eval_average_metrics": 73.9747073753716, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6756756756756757, \"recall\": 0.6944444444444444, \"f1-score\": 0.684931506849315, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.78125, \"recall\": 0.6410256410256411, \"f1-score\": 0.7042253521126761, \"support\": 39.0}, \"question\": {\"precision\": 0.7, \"recall\": 0.7777777777777778, \"f1-score\": 0.7368421052631577, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5833333333333334, \"recall\": 0.875, \"f1-score\": 0.7000000000000001, \"support\": 16.0}, \"accuracy\": 0.7333333333333333, \"macro avg\": {\"precision\": 0.8040708376645876, \"recall\": 0.7455686674436675, \"f1-score\": 0.7566759148589195, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7558651279484612, \"recall\": 0.7333333333333333, \"f1-score\": 0.7356457134892781, \"support\": 135.0}}", "eval_f1_macro": 75.66759148589195, "eval_f1_micro": 73.33333333333333, "eval_f1_weighted": 73.5645713489278, "eval_loss": 1.2941668033599854, "eval_runtime": 2.1293, "eval_samples_per_second": 63.402, "step": 2975 }, { "epoch": 176.0, "learning_rate": 3.9999999999999996e-05, "loss": 0.0002, "step": 2992 }, { "epoch": 176.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.61480524125062, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7105263157894737, \"recall\": 0.6923076923076923, \"f1-score\": 0.7012987012987013, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7935292547718269, \"recall\": 0.7337497571872572, \"f1-score\": 0.7460040211884412, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7400211319000227, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215511514245466, \"support\": 135.0}}", "eval_f1_macro": 74.60040211884412, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.15511514245466, "eval_loss": 1.3083090782165527, "eval_runtime": 2.2303, "eval_samples_per_second": 60.529, "step": 2992 }, { "epoch": 177.0, "learning_rate": 3.833333333333333e-05, "loss": 0.0002, "step": 3009 }, { "epoch": 177.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.60305283815627, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7928594842657343, \"recall\": 0.7294094794094794, \"f1-score\": 0.7450467847762234, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7376262626262626, \"recall\": 0.7185185185185186, \"f1-score\": 0.7220382917129899, \"support\": 135.0}}", "eval_f1_macro": 74.50467847762235, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.203829171299, "eval_loss": 1.3025906085968018, "eval_runtime": 2.2827, "eval_samples_per_second": 59.141, "step": 3009 }, { "epoch": 178.0, "learning_rate": 3.666666666666666e-05, "loss": 0.0001, "step": 3026 }, { "epoch": 178.0, "eval_accuracy": 72.5925925925926, "eval_average_metrics": 73.32833124918041, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.8181818181818182, \"recall\": 0.8181818181818182, \"f1-score\": 0.8181818181818182, \"support\": 11.0}, \"challenge\": {\"precision\": 0.71875, \"recall\": 0.6388888888888888, \"f1-score\": 0.676470588235294, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7, \"recall\": 0.717948717948718, \"f1-score\": 0.708860759493671, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5454545454545454, \"recall\": 0.75, \"f1-score\": 0.631578947368421, \"support\": 16.0}, \"accuracy\": 0.725925925925926, \"macro avg\": {\"precision\": 0.80234375, \"recall\": 0.7326146076146076, \"f1-score\": 0.7515625584906763, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7454040404040404, \"recall\": 0.725925925925926, \"f1-score\": 0.7297188396246881, \"support\": 135.0}}", "eval_f1_macro": 75.15625584906763, "eval_f1_micro": 72.59259259259261, "eval_f1_weighted": 72.97188396246881, "eval_loss": 1.2811334133148193, "eval_runtime": 2.3185, "eval_samples_per_second": 58.228, "step": 3026 }, { "epoch": 179.0, "learning_rate": 3.5e-05, "loss": 0.002, "step": 3043 }, { "epoch": 179.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.94066572336109, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.75, \"recall\": 0.8181818181818182, \"f1-score\": 0.7826086956521738, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6923076923076923, \"recall\": 0.6923076923076923, \"f1-score\": 0.6923076923076923, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7887609848076188, \"recall\": 0.7259372571872572, \"f1-score\": 0.740553221987642, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7323961919333588, \"recall\": 0.7111111111111111, \"f1-score\": 0.7148511847245799, \"support\": 135.0}}", "eval_f1_macro": 74.05532219876419, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.48511847245798, "eval_loss": 1.3122072219848633, "eval_runtime": 2.2232, "eval_samples_per_second": 60.724, "step": 3043 }, { "epoch": 180.0, "learning_rate": 3.333333333333333e-05, "loss": 0.0001, "step": 3060 }, { "epoch": 180.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.62519417356738, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", "eval_f1_macro": 74.72856934535828, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06850364520756, "eval_loss": 1.3237799406051636, "eval_runtime": 2.2054, "eval_samples_per_second": 61.213, "step": 3060 }, { "epoch": 181.0, "learning_rate": 3.1666666666666666e-05, "loss": 0.0001, "step": 3077 }, { "epoch": 181.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.62519417356738, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", "eval_f1_macro": 74.72856934535828, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06850364520756, "eval_loss": 1.3265109062194824, "eval_runtime": 2.3696, "eval_samples_per_second": 56.971, "step": 3077 }, { "epoch": 182.0, "learning_rate": 2.9999999999999997e-05, "loss": 0.0, "step": 3094 }, { "epoch": 182.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.62519417356738, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", "eval_f1_macro": 74.72856934535828, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06850364520756, "eval_loss": 1.3261035680770874, "eval_runtime": 2.2698, "eval_samples_per_second": 59.477, "step": 3094 }, { "epoch": 183.0, "learning_rate": 2.833333333333333e-05, "loss": 0.0009, "step": 3111 }, { "epoch": 183.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.22624100757768, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5217391304347826, \"recall\": 0.75, \"f1-score\": 0.6153846153846153, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.7805373005970924, \"recall\": 0.722732128982129, \"f1-score\": 0.7343542579082432, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7253561559459545, \"recall\": 0.7037037037037037, \"f1-score\": 0.7072879749874571, \"support\": 135.0}}", "eval_f1_macro": 73.43542579082431, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.7287974987457, "eval_loss": 1.2959227561950684, "eval_runtime": 2.1932, "eval_samples_per_second": 61.554, "step": 3111 }, { "epoch": 184.0, "learning_rate": 2.6666666666666667e-05, "loss": 0.0, "step": 3128 }, { "epoch": 184.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.55073789502072, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7272727272727273, \"recall\": 0.6666666666666666, \"f1-score\": 0.6956521739130435, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.6842105263157895, \"recall\": 0.6666666666666666, \"f1-score\": 0.6753246753246753, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7889478942110522, \"recall\": 0.7296765734265734, \"f1-score\": 0.7433792315665717, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7359373569899886, \"recall\": 0.7185185185185186, \"f1-score\": 0.7216132471972199, \"support\": 135.0}}", "eval_f1_macro": 74.33792315665717, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.161324719722, "eval_loss": 1.2773268222808838, "eval_runtime": 2.157, "eval_samples_per_second": 62.586, "step": 3128 }, { "epoch": 185.0, "learning_rate": 2.4999999999999998e-05, "loss": 0.0007, "step": 3145 }, { "epoch": 185.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.5453732812465, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7885856194679723, \"recall\": 0.7296765734265734, \"f1-score\": 0.7432477269156164, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7355754414577943, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215301672972063, \"support\": 135.0}}", "eval_f1_macro": 74.32477269156163, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.15301672972065, "eval_loss": 1.2745593786239624, "eval_runtime": 2.2478, "eval_samples_per_second": 60.059, "step": 3145 }, { "epoch": 186.0, "learning_rate": 2.3333333333333332e-05, "loss": 0.0001, "step": 3162 }, { "epoch": 186.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.5453732812465, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7058823529411765, \"recall\": 0.6666666666666666, \"f1-score\": 0.6857142857142857, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5714285714285714, \"recall\": 0.75, \"f1-score\": 0.6486486486486486, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7885856194679723, \"recall\": 0.7296765734265734, \"f1-score\": 0.7432477269156164, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7355754414577943, \"recall\": 0.7185185185185186, \"f1-score\": 0.7215301672972063, \"support\": 135.0}}", "eval_f1_macro": 74.32477269156163, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.15301672972065, "eval_loss": 1.273395299911499, "eval_runtime": 2.1937, "eval_samples_per_second": 61.539, "step": 3162 }, { "epoch": 187.0, "learning_rate": 2.1666666666666664e-05, "loss": 0.002, "step": 3179 }, { "epoch": 187.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89929243227385, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.785339764674442, \"recall\": 0.730544628982129, \"f1-score\": 0.7397919123590555, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7330601223074342, \"recall\": 0.7111111111111111, \"f1-score\": 0.7139575627096765, \"support\": 135.0}}", "eval_f1_macro": 73.97919123590555, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.39575627096765, "eval_loss": 1.2829253673553467, "eval_runtime": 2.2014, "eval_samples_per_second": 61.325, "step": 3179 }, { "epoch": 188.0, "learning_rate": 1.9999999999999998e-05, "loss": 0.0001, "step": 3196 }, { "epoch": 188.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.62519417356738, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7333333333333333, \"recall\": 0.6111111111111112, \"f1-score\": 0.6666666666666666, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7027027027027027, \"recall\": 0.6666666666666666, \"f1-score\": 0.6842105263157895, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7910440066690066, \"recall\": 0.7419082653457654, \"f1-score\": 0.7472856934535829, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7411591678258346, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206850364520756, \"support\": 135.0}}", "eval_f1_macro": 74.72856934535828, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06850364520756, "eval_loss": 1.3078842163085938, "eval_runtime": 2.2833, "eval_samples_per_second": 59.124, "step": 3196 }, { "epoch": 189.0, "learning_rate": 1.833333333333333e-05, "loss": 0.0, "step": 3213 }, { "epoch": 189.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.62216036599705, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}", "eval_f1_macro": 74.71822632362485, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06671143665966, "eval_loss": 1.2980728149414062, "eval_runtime": 2.2356, "eval_samples_per_second": 60.387, "step": 3213 }, { "epoch": 190.0, "learning_rate": 1.6666666666666664e-05, "loss": 0.0, "step": 3230 }, { "epoch": 190.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.62216036599705, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}", "eval_f1_macro": 74.71822632362485, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06671143665966, "eval_loss": 1.2942575216293335, "eval_runtime": 2.1696, "eval_samples_per_second": 62.223, "step": 3230 }, { "epoch": 191.0, "learning_rate": 1.4999999999999999e-05, "loss": 0.0001, "step": 3247 }, { "epoch": 191.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.974518342938, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7867627164502164, \"recall\": 0.7387031371406372, \"f1-score\": 0.7437678567314118, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7322831489498156, \"recall\": 0.7111111111111111, \"f1-score\": 0.7129906547638859, \"support\": 135.0}}", "eval_f1_macro": 74.37678567314117, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.29906547638859, "eval_loss": 1.293574571609497, "eval_runtime": 2.1704, "eval_samples_per_second": 62.2, "step": 3247 }, { "epoch": 192.0, "learning_rate": 1.3333333333333333e-05, "loss": 0.0, "step": 3264 }, { "epoch": 192.0, "eval_accuracy": 71.85185185185186, "eval_average_metrics": 72.62216036599705, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7142857142857143, \"recall\": 0.9090909090909091, \"f1-score\": 0.8, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7096774193548387, \"recall\": 0.6111111111111112, \"f1-score\": 0.6567164179104478, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7185185185185186, \"macro avg\": {\"precision\": 0.7905269573616347, \"recall\": 0.7419082653457654, \"f1-score\": 0.7471822632362485, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.740489896403875, \"recall\": 0.7185185185185186, \"f1-score\": 0.7206671143665966, \"support\": 135.0}}", "eval_f1_macro": 74.71822632362485, "eval_f1_micro": 71.85185185185186, "eval_f1_weighted": 72.06671143665966, "eval_loss": 1.2941248416900635, "eval_runtime": 2.1999, "eval_samples_per_second": 61.367, "step": 3264 }, { "epoch": 193.0, "learning_rate": 1.1666666666666666e-05, "loss": 0.0001, "step": 3281 }, { "epoch": 193.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89912419879386, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7850075271950272, \"recall\": 0.730544628982129, \"f1-score\": 0.739725063938619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327851161184494, \"recall\": 0.7111111111111111, \"f1-score\": 0.714017681790913, \"support\": 135.0}}", "eval_f1_macro": 73.9725063938619, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.40176817909129, "eval_loss": 1.2996220588684082, "eval_runtime": 2.3185, "eval_samples_per_second": 58.228, "step": 3281 }, { "epoch": 194.0, "learning_rate": 9.999999999999999e-06, "loss": 0.0001, "step": 3298 }, { "epoch": 194.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.89912419879386, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6875, \"recall\": 0.6111111111111112, \"f1-score\": 0.6470588235294118, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7222222222222222, \"recall\": 0.6666666666666666, \"f1-score\": 0.6933333333333334, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7850075271950272, \"recall\": 0.730544628982129, \"f1-score\": 0.739725063938619, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7327851161184494, \"recall\": 0.7111111111111111, \"f1-score\": 0.714017681790913, \"support\": 135.0}}", "eval_f1_macro": 73.9725063938619, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.40176817909129, "eval_loss": 1.3008208274841309, "eval_runtime": 2.2664, "eval_samples_per_second": 59.567, "step": 3298 }, { "epoch": 195.0, "learning_rate": 8.333333333333332e-06, "loss": 0.0001, "step": 3315 }, { "epoch": 195.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.2542231596473, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.781411297036297, \"recall\": 0.7273395007770008, \"f1-score\": 0.7363456487177715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7249367916034583, \"recall\": 0.7037037037037037, \"f1-score\": 0.7064158702607126, \"support\": 135.0}}", "eval_f1_macro": 73.63456487177716, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.64158702607126, "eval_loss": 1.2944679260253906, "eval_runtime": 2.2091, "eval_samples_per_second": 61.112, "step": 3315 }, { "epoch": 196.0, "learning_rate": 6.666666666666667e-06, "loss": 0.0154, "step": 3332 }, { "epoch": 196.0, "eval_accuracy": 70.37037037037037, "eval_average_metrics": 71.2542231596473, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6666666666666666, \"recall\": 0.6111111111111112, \"f1-score\": 0.6376811594202899, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5416666666666666, \"recall\": 0.8125, \"f1-score\": 0.65, \"support\": 16.0}, \"accuracy\": 0.7037037037037037, \"macro avg\": {\"precision\": 0.781411297036297, \"recall\": 0.7273395007770008, \"f1-score\": 0.7363456487177715, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7249367916034583, \"recall\": 0.7037037037037037, \"f1-score\": 0.7064158702607126, \"support\": 135.0}}", "eval_f1_macro": 73.63456487177716, "eval_f1_micro": 70.37037037037037, "eval_f1_weighted": 70.64158702607126, "eval_loss": 1.2910572290420532, "eval_runtime": 2.3432, "eval_samples_per_second": 57.614, "step": 3332 }, { "epoch": 197.0, "learning_rate": 4.9999999999999996e-06, "loss": 0.0003, "step": 3349 }, { "epoch": 197.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.91662203626719, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}", "eval_f1_macro": 74.08616942664258, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.35809649620393, "eval_loss": 1.287298560142517, "eval_runtime": 2.18, "eval_samples_per_second": 61.926, "step": 3349 }, { "epoch": 198.0, "learning_rate": 3.3333333333333333e-06, "loss": 0.0001, "step": 3366 }, { "epoch": 198.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.91662203626719, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}", "eval_f1_macro": 74.08616942664258, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.35809649620393, "eval_loss": 1.2868810892105103, "eval_runtime": 2.1185, "eval_samples_per_second": 63.724, "step": 3366 }, { "epoch": 199.0, "learning_rate": 1.6666666666666667e-06, "loss": 0.0002, "step": 3383 }, { "epoch": 199.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.91662203626719, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}", "eval_f1_macro": 74.08616942664258, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.35809649620393, "eval_loss": 1.2866387367248535, "eval_runtime": 2.1869, "eval_samples_per_second": 61.732, "step": 3383 }, { "epoch": 200.0, "learning_rate": 0.0, "loss": 0.0081, "step": 3400 }, { "epoch": 200.0, "eval_accuracy": 71.11111111111111, "eval_average_metrics": 71.91662203626719, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.6923076923076923, \"recall\": 0.8181818181818182, \"f1-score\": 0.7500000000000001, \"support\": 11.0}, \"challenge\": {\"precision\": 0.6764705882352942, \"recall\": 0.6388888888888888, \"f1-score\": 0.6571428571428571, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7142857142857143, \"recall\": 0.6410256410256411, \"f1-score\": 0.6756756756756757, \"support\": 39.0}, \"question\": {\"precision\": 0.6363636363636364, \"recall\": 0.7777777777777778, \"f1-score\": 0.7000000000000001, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.7692307692307693, \"f1-score\": 0.8695652173913044, \"support\": 13.0}, \"request\": {\"precision\": 0.5652173913043478, \"recall\": 0.8125, \"f1-score\": 0.6666666666666667, \"support\": 16.0}, \"accuracy\": 0.7111111111111111, \"macro avg\": {\"precision\": 0.7855806278120856, \"recall\": 0.730811722999223, \"f1-score\": 0.7408616942664257, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.7303423676825213, \"recall\": 0.7111111111111111, \"f1-score\": 0.7135809649620393, \"support\": 135.0}}", "eval_f1_macro": 74.08616942664258, "eval_f1_micro": 71.11111111111111, "eval_f1_weighted": 71.35809649620393, "eval_loss": 1.2865550518035889, "eval_runtime": 2.1903, "eval_samples_per_second": 61.635, "step": 3400 }, { "epoch": 200.0, "step": 3400, "total_flos": 1.9410334099963904e+16, "train_runtime": 2245.6861, "train_samples_per_second": 1.514 } ], "max_steps": 3400, "num_train_epochs": 200, "total_flos": 1.9410334099963904e+16, "trial_name": null, "trial_params": null }