{ "best_metric": 1.147840142250061, "best_model_checkpoint": "./mistral/10-04-24-Weni-WeniGPT-Agents-Mistral-1.0.0-SFT_Experiment with SFT and a new tokenizer configuration for chat template of mistral-2_max_steps-312_batch_4_2024-04-10_ppid_9/checkpoint-100", "epoch": 1.9138755980861244, "eval_steps": 50, "global_step": 200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.19, "grad_norm": 0.6767070889472961, "learning_rate": 0.0001993503206718859, "loss": 1.421, "step": 20 }, { "epoch": 0.38, "grad_norm": 0.7594590783119202, "learning_rate": 0.00019487887022684336, "loss": 1.1094, "step": 40 }, { "epoch": 0.48, "eval_loss": 1.1684353351593018, "eval_runtime": 15.9813, "eval_samples_per_second": 2.878, "eval_steps_per_second": 1.439, "step": 50 }, { "epoch": 0.57, "grad_norm": 0.8954619765281677, "learning_rate": 0.00018634217048966637, "loss": 1.0392, "step": 60 }, { "epoch": 0.77, "grad_norm": 0.6521860361099243, "learning_rate": 0.0001741059911251997, "loss": 0.9187, "step": 80 }, { "epoch": 0.96, "grad_norm": 0.869737446308136, "learning_rate": 0.0001586946124013354, "loss": 0.879, "step": 100 }, { "epoch": 0.96, "eval_loss": 1.147840142250061, "eval_runtime": 15.9788, "eval_samples_per_second": 2.879, "eval_steps_per_second": 1.439, "step": 100 }, { "epoch": 1.15, "grad_norm": 0.7937932014465332, "learning_rate": 0.00014076836149416887, "loss": 0.6586, "step": 120 }, { "epoch": 1.34, "grad_norm": 1.0015738010406494, "learning_rate": 0.00012109531962807332, "loss": 0.6885, "step": 140 }, { "epoch": 1.44, "eval_loss": 1.1584892272949219, "eval_runtime": 15.9777, "eval_samples_per_second": 2.879, "eval_steps_per_second": 1.44, "step": 150 }, { "epoch": 1.53, "grad_norm": 0.8210965991020203, "learning_rate": 0.00010051841230721065, "loss": 0.6127, "step": 160 }, { "epoch": 1.72, "grad_norm": 0.9020585417747498, "learning_rate": 7.991929271442817e-05, "loss": 0.4806, "step": 180 }, { "epoch": 1.91, "grad_norm": 1.1187869310379028, "learning_rate": 6.018056575578075e-05, "loss": 0.5073, "step": 200 }, { "epoch": 1.91, "eval_loss": 1.1682521104812622, "eval_runtime": 15.98, "eval_samples_per_second": 2.879, "eval_steps_per_second": 1.439, "step": 200 } ], "logging_steps": 20, "max_steps": 312, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "total_flos": 1.004587688117207e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }