{ "epoch": 200.0, "eval_accuracy": 77.03703703703704, "eval_average_metrics": 77.53590008051296, "eval_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7941176470588235, \"recall\": 0.6923076923076923, \"f1-score\": 0.7397260273972601, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7703703703703704, \"macro avg\": {\"precision\": 0.833837516113058, \"recall\": 0.7720898892773893, \"f1-score\": 0.7884724218577965, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.784561747926041, \"recall\": 0.7703703703703704, \"f1-score\": 0.7722228406219811, \"support\": 135.0}}", "eval_f1_macro": 78.84724218577965, "eval_f1_micro": 77.03703703703704, "eval_f1_weighted": 77.22228406219811, "eval_loss": 1.0223504304885864, "eval_runtime": 1.7125, "eval_samples_per_second": 78.832, "init_mem_cpu_alloc_delta": -543014912, "init_mem_cpu_peaked_delta": 543023104, "init_mem_gpu_alloc_delta": 891528192, "init_mem_gpu_peaked_delta": 0, "peak_memory": 5.305544921875, "test_accuracy": 77.03703703703704, "test_average_metrics": 77.53590008051296, "test_classification_report": "{\"acceptance\": {\"precision\": 1.0, \"recall\": 0.8888888888888888, \"f1-score\": 0.9411764705882353, \"support\": 9.0}, \"apology\": {\"precision\": 1.0, \"recall\": 0.5, \"f1-score\": 0.6666666666666666, \"support\": 2.0}, \"appreciation\": {\"precision\": 0.7692307692307693, \"recall\": 0.9090909090909091, \"f1-score\": 0.8333333333333333, \"support\": 11.0}, \"challenge\": {\"precision\": 0.7105263157894737, \"recall\": 0.75, \"f1-score\": 0.7297297297297298, \"support\": 36.0}, \"informing statement\": {\"precision\": 0.7941176470588235, \"recall\": 0.6923076923076923, \"f1-score\": 0.7397260273972601, \"support\": 39.0}, \"question\": {\"precision\": 0.7777777777777778, \"recall\": 0.7777777777777778, \"f1-score\": 0.7777777777777778, \"support\": 9.0}, \"rejection\": {\"precision\": 1.0, \"recall\": 0.8461538461538461, \"f1-score\": 0.9166666666666666, \"support\": 13.0}, \"request\": {\"precision\": 0.6190476190476191, \"recall\": 0.8125, \"f1-score\": 0.7027027027027026, \"support\": 16.0}, \"accuracy\": 0.7703703703703704, \"macro avg\": {\"precision\": 0.833837516113058, \"recall\": 0.7720898892773893, \"f1-score\": 0.7884724218577965, \"support\": 135.0}, \"weighted avg\": {\"precision\": 0.784561747926041, \"recall\": 0.7703703703703704, \"f1-score\": 0.7722228406219811, \"support\": 135.0}}", "test_f1_macro": 78.84724218577965, "test_f1_micro": 77.03703703703704, "test_f1_weighted": 77.22228406219811, "test_loss": 1.0223504304885864, "test_runtime": 1.6896, "test_samples_per_second": 79.902, "total_time in minutes ": 37.44455833333333, "train_mem_cpu_alloc_delta": 1023647744, "train_mem_cpu_peaked_delta": 32768, "train_mem_gpu_alloc_delta": 40251392, "train_mem_gpu_peaked_delta": 4624237568, "train_runtime": 2245.6861, "train_samples": 541, "train_samples_per_second": 1.514 }