TheBloke commited on
Commit
87924a5
1 Parent(s): c3cfb5a

Initial GPTQ model commit.

Browse files
Files changed (1) hide show
  1. README.md +8 -6
README.md CHANGED
@@ -5,17 +5,19 @@ datasets:
5
  inference: false
6
  ---
7
 
8
- # WizardLM 30B Uncensored
9
 
10
- These files are GPTQ 4bit model files for [Eric Hartford's WizardLM 30B 'uncensored'](https://huggingface.co/ehartford/WizardLM-30B-Uncensored).
11
 
12
  It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
13
 
 
 
14
  ## Other repositories available
15
 
16
- * [4bit GPTQ model for GPU inference](https://huggingface.co/TheBloke/WizardLM-30B-Uncensored-GPTQ)
17
- * [4-bit, 5-bit and 8-bit GGML models for CPU (+CUDA) inference](https://huggingface.co/TheBloke/WizardLM-30B-Uncensored-GGML)
18
- * [Eric's unquantised model in fp16 HF format](https://huggingface.co/ehartford/WizardLM-30B-Uncensored)
19
 
20
  ## How to easily download and use this model in text-generation-webui
21
 
@@ -43,7 +45,7 @@ This will work with all versions of GPTQ-for-LLaMa. It has maximum compatibility
43
 
44
  It was created without the `--act-order` parameter. It may have slightly lower inference quality compared to the other file, but is guaranteed to work on all versions of GPTQ-for-LLaMa and text-generation-webui.
45
 
46
- * `WizardLM-30B-uncensored-GPTQ-4bit.act-order.safetensors`
47
  * Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
48
  * Works with AutoGPTQ. Use `strict=False` to load.
49
  * Works with text-generation-webui one-click-installers
 
5
  inference: false
6
  ---
7
 
8
+ # WizardLM - uncensored: An Instruction-following LLM Using Evol-Instruct
9
 
10
+ These files are GPTQ 4bit model files for [Eric Hartford's 'uncensored' version of WizardLM](https://huggingface.co/ehartford/WizardLM-30B-Uncensored).
11
 
12
  It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa).
13
 
14
+ Eric did a fresh 7B training using the WizardLM method, on [a dataset edited to remove all the "I'm sorry.." type ChatGPT responses](https://huggingface.co/datasets/ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered).
15
+
16
  ## Other repositories available
17
 
18
+ * [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/WizardLM-30B-uncensored-GPTQ)
19
+ * [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/WizardLM-30B-uncensored-GGML)
20
+ * [Eric's unquantised model in HF format](https://huggingface.co/ehartford/WizardLM-30B-Uncensored)
21
 
22
  ## How to easily download and use this model in text-generation-webui
23
 
 
45
 
46
  It was created without the `--act-order` parameter. It may have slightly lower inference quality compared to the other file, but is guaranteed to work on all versions of GPTQ-for-LLaMa and text-generation-webui.
47
 
48
+ * `wizard-vicuna-13B-GPTQ-4bit.compat.no-act-order.safetensors`
49
  * Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
50
  * Works with AutoGPTQ. Use `strict=False` to load.
51
  * Works with text-generation-webui one-click-installers