GGUF llama.cpp quantized version of:
https://huggingface.co/blog/llama2#how-to-prompt-llama-2
2-bit
4-bit
5-bit
8-bit