LLaMA 7B - AWQ GGUF
Description
This repo contains GGUF format model files with AWQ quantizaton support for Meta's LLaMA 7b.
About AWQ-GGUF
The model was convert by the combination of llama.cpp and quantization method AWQ
How to use models
Please refer to the PR from llama.cpp for more detail.