SLM_vs_LLM_experiments
/
max_seq_length_128_experiments
/google_bert
/bert_base_uncased_amazon
/train_results.json
{ | |
"epoch": 3.0, | |
"train_loss": 1.1961698414986595, | |
"train_runtime": 135.7278, | |
"train_samples": 12144, | |
"train_samples_per_second": 268.42, | |
"train_steps_per_second": 4.2 | |
} |