TheBloke commited on
Commit
f80b1d9
1 Parent(s): 3060ef2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -0
README.md CHANGED
@@ -7,6 +7,20 @@ This is an HF version of the [Vicuna 7B 1.1 model](https://huggingface.co/lmsys/
7
 
8
  It was created by merging the deltas provided in the above repo with the original Llama 7B model, [using the code provided on their Github page](https://github.com/lm-sys/FastChat#vicuna-weights).
9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  # Vicuna Model Card
11
 
12
  ## Model details
 
7
 
8
  It was created by merging the deltas provided in the above repo with the original Llama 7B model, [using the code provided on their Github page](https://github.com/lm-sys/FastChat#vicuna-weights).
9
 
10
+ ## My Vicuna 1.1 model repositories
11
+
12
+ I have the following Vicuna 1.1 repositories available:
13
+
14
+ **13B models:**
15
+ * [Unquantized 13B 1.1 model for GPU - HF format](https://huggingface.co/TheBloke/vicuna-13B-1.1-HF)
16
+ * [GPTQ quantized 4bit 13B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g)
17
+ * [GPTQ quantized 4bit 13B 1.1 for CPU - GGML format for `llama.cpp`](https://huggingface.co/TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g-GGML)
18
+
19
+ **7B models:**
20
+ * [Unquantized 7B 1.1 model for GPU - HF format](https://huggingface.co/TheBloke/vicuna-7B-1.1-HF)
21
+ * [GPTQ quantized 4bit 7B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g)
22
+ * [GPTQ quantized 4bit 7B 1.1 for CPU - GGML format for `llama.cpp`](https://huggingface.co/TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g-GGML)
23
+
24
  # Vicuna Model Card
25
 
26
  ## Model details