TheBloke commited on
Commit
065ee41
1 Parent(s): ad48bb6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -23,13 +23,13 @@ I have the following Koala model repositories available:
23
  **13B models:**
24
  * [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
25
  * [GPTQ quantized 4bit 13B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g)
26
- * [GPTQ quantized 4bit 13B model in GGML format for `llama.cpp`](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g-GGML)
27
 
28
  **7B models:**
29
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
30
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
31
  * [GPTQ quantized 4bit 7B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-7B-GPTQ-4bit-128g)
32
- * [GPTQ quantized 4bit 7B model in GGML format for `llama.cpp`](https://huggingface.co/TheBloke/koala-7B-GPTQ-4bit-128g-GGML)
33
 
34
  ## GETTING GIBBERISH OUTPUT?
35
 
 
23
  **13B models:**
24
  * [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
25
  * [GPTQ quantized 4bit 13B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-13B-GPTQ-4bit-128g)
26
+ * [4-bit, 5-bit and 8-bit GGML models for `llama.cpp`](https://huggingface.co/TheBloke/koala-13B-GGML)
27
 
28
  **7B models:**
29
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
30
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
31
  * [GPTQ quantized 4bit 7B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-7B-GPTQ-4bit-128g)
32
+ * [4-bit, 5-bit and 8-bit GGML models for `llama.cpp`](https://huggingface.co/TheBloke/koala-7B-GGML)
33
 
34
  ## GETTING GIBBERISH OUTPUT?
35