This is a Llama-2-7b-chat-hf model fine-tuned using QLoRA (4-bit precision) on the mlabonne/guanaco-llama2-1k dataset(a subset of OpenAssistant/oasst1).
This model was trained on a single A100.
{'train_runtime': 2368.6484, 'train_samples_per_second': 4.157, 'train_steps_per_second': 1.039, 'train_loss': 1.3207073250405679, 'epoch': 1.0}
This model was finetuned primarily for educational purposes.
- Downloads last month
- 21
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.