gpt-for-est-large / all_results.json
mphi's picture
Upload all_results.json
817bcfc
raw
history blame
403 Bytes
{
"epoch": 3.0,
"eval_loss": 3.522364616394043,
"eval_runtime": 456.0188,
"eval_samples": 32253,
"eval_samples_per_second": 70.727,
"eval_steps_per_second": 4.421,
"perplexity": 33.86441020305686,
"train_loss": 1.773998625050055,
"train_runtime": 174345.3711,
"train_samples": 3196485,
"train_samples_per_second": 55.003,
"train_steps_per_second": 1.719
}