TheBloke commited on
Commit
d2630c0
1 Parent(s): 0b8949b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -30,7 +30,7 @@ GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/gger
30
 
31
  * [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GPTQ)
32
  * [4-bit, 5-bit, and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GGML)
33
- * [Original unquantised fp16 model in HF format](https://huggingface.co/TheBloke/gorilla-7B-HF)
34
 
35
  ## THE FILES IN MAIN BRANCH REQUIRES LATEST LLAMA.CPP (May 19th 2023 - commit 2d5db48)!
36
 
 
30
 
31
  * [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GPTQ)
32
  * [4-bit, 5-bit, and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GGML)
33
+ * [Merged, unquantised fp16 model in HF format](https://huggingface.co/TheBloke/gorilla-7B-fp16)
34
 
35
  ## THE FILES IN MAIN BRANCH REQUIRES LATEST LLAMA.CPP (May 19th 2023 - commit 2d5db48)!
36