SLM_vs_LLM_experiments
/
max_seq_length_128_experiments
/google_t5
/t5_small_amazon
/all_results.json
{ | |
"epoch": 3.0, | |
"eval_accuracy": 0.7766798418972332, | |
"eval_f1_macro": 0.7273078975523218, | |
"eval_f1_micro": 0.7766798418972332, | |
"eval_loss": 0.7013603448867798, | |
"eval_runtime": 1.2598, | |
"eval_samples": 1518, | |
"eval_samples_per_second": 1204.953, | |
"eval_steps_per_second": 38.101, | |
"test_accuracy": 0.8063241106719368, | |
"test_f1_macro": 0.7499845618089623, | |
"test_f1_micro": 0.8063241106719368, | |
"test_loss": 0.6868003606796265, | |
"test_runtime": 1.3109, | |
"test_samples_per_second": 1157.949, | |
"test_steps_per_second": 36.615, | |
"train_loss": 0.8189868109268055, | |
"train_runtime": 138.2392, | |
"train_samples": 12144, | |
"train_samples_per_second": 263.543, | |
"train_steps_per_second": 8.247 | |
} |