eachadea's picture
Update README.md
6e50f34
|
raw
history blame
455 Bytes
---
pipeline_tag: conversational
tags:
- vicuna
- llama
- text-generation-inference
---
Converted for use with [llama.cpp](https://github.com/ggerganov/llama.cpp)
---
- 4-bit quantized
- Needs ~10GB of CPU RAM
- Won't work with alpaca.cpp or old llama.cpp (new ggml format requires latest llama.cpp)
---
Smaller 7B version can be found here: https://huggingface.co/eachadea/ggml-vicuna-7b-4bit
---
tags:
- vicuna
- llama
- text-generation-inference
---