SLM_vs_LLM_experiments
/
max_seq_length_128_experiments
/google_t5
/t5_base_twitter
/all_results.json
{ | |
"epoch": 3.0, | |
"eval_accuracy": 0.765625, | |
"eval_f1_macro": 0.7266124240384777, | |
"eval_f1_micro": 0.765625, | |
"eval_loss": 0.49125435948371887, | |
"eval_runtime": 2.6918, | |
"eval_samples": 1088, | |
"eval_samples_per_second": 404.194, | |
"eval_steps_per_second": 12.631, | |
"test_accuracy": 0.7738970588235294, | |
"test_f1_macro": 0.7347446036748528, | |
"test_f1_micro": 0.7738970588235294, | |
"test_loss": 0.4866201877593994, | |
"test_runtime": 2.6974, | |
"test_samples_per_second": 403.349, | |
"test_steps_per_second": 12.605, | |
"train_loss": 0.3840603100902894, | |
"train_runtime": 309.7161, | |
"train_samples": 8700, | |
"train_samples_per_second": 84.271, | |
"train_steps_per_second": 2.635 | |
} |