Edit model card

This is a Llama-2-7b-chat-hf model fine-tuned using QLoRA (4-bit precision) on the mlabonne/guanaco-llama2-1k dataset(a subset of OpenAssistant/oasst1).

This model was trained on a single A100.

{'train_runtime': 2368.6484, 'train_samples_per_second': 4.157, 'train_steps_per_second': 1.039, 'train_loss': 1.3207073250405679, 'epoch': 1.0}

This model was finetuned primarily for educational purposes.

Downloads last month
2

Dataset used to train aloobun/llama2-7b-guanaco