SLM_vs_LLM_experiments
/
max_seq_length_128_experiments
/Qwen
/Qwen1.5_1.8B_amazon
/all_results.json
{ | |
"epoch": 1.0, | |
"eval_accuracy": 0.852437417654809, | |
"eval_f1_macro": 0.8289709215944268, | |
"eval_f1_micro": 0.852437417654809, | |
"eval_loss": 0.5073133111000061, | |
"eval_runtime": 3.9722, | |
"eval_samples": 1518, | |
"eval_samples_per_second": 382.155, | |
"eval_steps_per_second": 12.084, | |
"test_accuracy": 0.8656126482213439, | |
"test_f1_macro": 0.8295835523548035, | |
"test_f1_micro": 0.8656126482213439, | |
"test_loss": 0.4794342815876007, | |
"test_runtime": 4.0382, | |
"test_samples_per_second": 375.908, | |
"test_steps_per_second": 11.886, | |
"train_loss": 1.0740083393297697, | |
"train_runtime": 415.974, | |
"train_samples": 12144, | |
"train_samples_per_second": 29.194, | |
"train_steps_per_second": 0.914 | |
} |