File size: 514 Bytes
bb46359
 
 
 
 
 
 
 
 
40313ef
bb46359
77ba78a
bb46359
4d101c5
 
 
 
 
bb46359
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
---
pipeline_tag: conversational
tags:
- vicuna
- llama
- text-generation-inference
---
Converted for use with [llama.cpp](https://github.com/ggerganov/llama.cpp)
---
- Based on AlekseyKorshuk/vicuna-7b
- 4-bit quantized
- Needs ~6GB of CPU RAM
- Won't work with alpaca.cpp or old llama.cpp (new ggml format requires latest llama.cpp)
- 7B parameter version

---

Bigger 13B version can be found here: https://huggingface.co/eachadea/ggml-vicuna-13b-4bit

---
tags:
- vicuna
- llama
- text-generation-inference
---