TheBloke commited on
Commit
167767f
·
1 Parent(s): 8b3c6b2

Initial GPTQ model commit

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -19,7 +19,7 @@ model_type: llama
19
  <!-- header end -->
20
 
21
  # Llama-2-7B-Chat Code Cherry Pop - GPTQ
22
- - Model creator: TokenBender
23
  - Original model: [Llama-2-7B-Chat Code Cherry Pop](https://huggingface.co/TokenBender/llama2-7b-chat-hf-codeCherryPop-qLoRA-merged)
24
 
25
  ## Description
@@ -33,7 +33,7 @@ Multiple GPTQ parameter permutations are provided; see Provided Files below for
33
 
34
  * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/llama2-7b-chat-codeCherryPop-qLoRA-GPTQ)
35
  * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/llama2-7b-chat-codeCherryPop-qLoRA-GGML)
36
- * [TokenBender's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TokenBender/llama2-7b-chat-hf-codeCherryPop-qLoRA-merged)
37
 
38
  ## Prompt template: Alpaca
39
 
 
19
  <!-- header end -->
20
 
21
  # Llama-2-7B-Chat Code Cherry Pop - GPTQ
22
+ - Model creator: [TokenBender](https://huggingface.co/TokenBender)
23
  - Original model: [Llama-2-7B-Chat Code Cherry Pop](https://huggingface.co/TokenBender/llama2-7b-chat-hf-codeCherryPop-qLoRA-merged)
24
 
25
  ## Description
 
33
 
34
  * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/llama2-7b-chat-codeCherryPop-qLoRA-GPTQ)
35
  * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/llama2-7b-chat-codeCherryPop-qLoRA-GGML)
36
+ * [[TokenBender](https://huggingface.co/TokenBender)'s original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TokenBender/llama2-7b-chat-hf-codeCherryPop-qLoRA-merged)
37
 
38
  ## Prompt template: Alpaca
39