Edit model card

GGML/GGUF(v2) Quantizations of the model: https://huggingface.co/winglian/basilisk-4b This is winglian/llama-2-4b, a 4B parameter Llama-2 model, finetuned with open orca CoT data.

I tried to run on latest llama.cpp commit, but I was getting an error(GGML_ASSERT: llama.cpp:8136: false), then I converted again the model to gguf using this llama.cpp commit https://github.com/ggerganov/llama.cpp/tree/019ba1dcd0c7775a5ac0f7442634a330eb0173cc it seems to be working now.

Downloads last month
229
GGUF
Model size
3.5B params
Architecture
llama

Dataset used to train Aryanne/Basilisk-4B-gguf