LLaMA-7b-AWQ-GGUF / README.md
namtran's picture
update readme
dd6dd73
|
raw
history blame
654 Bytes

LLaMA 7B - AWQ GGUF

Description

This repo contains GGUF format model files with AWQ quantizaton support for Meta's LLaMA 7b.

About AWQ-GGUF

The model was convert by the combination of llama.cpp and quantization method AWQ

How to use models

Please refer to the PR from llama.cpp for more detail.