{ "base_model_name": "llama-2-70b-Guanaco-QLoRA-fp16", "base_model_class": "LlamaForCausalLM", "base_loaded_in_4bit": true, "base_loaded_in_8bit": false, "projections": "q, v", "loss": 0.964, "learning_rate": 8.68421052631579e-05, "epoch": 2.84, "current_steps": 1323, "train_runtime": 6257.8339, "train_samples_per_second": 0.908, "train_steps_per_second": 0.007, "total_flos": 2.8663842040971264e+17, "train_loss": 1.2440977039791288 }