SLM_vs_LLM_experiments
/
max_seq_length_128_experiments
/google_t5
/t5_base_amazon
/all_results.json
{ | |
"epoch": 3.0, | |
"eval_accuracy": 0.8399209486166008, | |
"eval_f1_macro": 0.8112898753723374, | |
"eval_f1_micro": 0.8399209486166008, | |
"eval_loss": 0.5565423369407654, | |
"eval_runtime": 3.8107, | |
"eval_samples": 1518, | |
"eval_samples_per_second": 398.35, | |
"eval_steps_per_second": 12.596, | |
"test_accuracy": 0.857707509881423, | |
"test_f1_macro": 0.8293055942786999, | |
"test_f1_micro": 0.857707509881423, | |
"test_loss": 0.5306673645973206, | |
"test_runtime": 3.7633, | |
"test_samples_per_second": 403.37, | |
"test_steps_per_second": 12.755, | |
"train_loss": 0.5285837122222834, | |
"train_runtime": 447.0645, | |
"train_samples": 12144, | |
"train_samples_per_second": 81.492, | |
"train_steps_per_second": 2.55 | |
} |