Meta's Llama-3 8B parameter base model trained on Alpaca dataset and outputed to 16bit GGUF Instruct model. Below is code for inference at command line using llama.cpp


./build/bin/main -m ./models/llama3_alpaca_dpo_GGUF-unsloth.F16.gguf \
-p '''Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n\n
### Instruction:\nWhy is the sky blue?\n\n
### Input:\n\n\n
### Response:\n'''

Downloads last month
2
GGUF
Model size
8.03B params
Architecture
llama

16-bit

Inference API
Unable to determine this model's library. Check the docs .