Llama-2-13b-alpaca-spanish-LoRA / training_log.json
marianbasti's picture
Upload files
d44d431
raw
history blame
458 Bytes
{
"base_model_name": "Llama-2-13B-fp16-padded",
"base_model_class": "LlamaForCausalLM",
"base_loaded_in_4bit": true,
"base_loaded_in_8bit": false,
"projections": "q, v",
"loss": 1.07,
"learning_rate": 1.7080653142008908e-05,
"epoch": 0.74,
"current_steps": 9678,
"train_runtime": 20262.2951,
"train_samples_per_second": 12.817,
"train_steps_per_second": 0.1,
"total_flos": 4.119249176808653e+17,
"train_loss": 1.1322209582423532
}