SLM_vs_LLM_experiments
/
max_seq_length_128_experiments
/LoRA
/Qwen
/Qwen1.5_7B_LoRA_MAdAiLab
/twitter_disaster
/train_results.json
{ | |
"epoch": 3.0, | |
"train_loss": 0.4493609409706265, | |
"train_runtime": 1238.3624, | |
"train_samples": 8700, | |
"train_samples_per_second": 21.076, | |
"train_steps_per_second": 0.659 | |
} |