TheBloke commited on
Commit
7bf6021
1 Parent(s): 4e67230

Updating model files

Browse files
Files changed (1) hide show
  1. README.md +25 -3
README.md CHANGED
@@ -6,6 +6,17 @@ tags:
6
  - llama
7
  inference: false
8
  ---
 
 
 
 
 
 
 
 
 
 
 
9
 
10
  # Wizard-Vicuna-13B-GPTQ
11
 
@@ -18,7 +29,7 @@ It is the result of quantising to 4bit using [GPTQ-for-LLaMa](https://github.com
18
  * [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-GPTQ).
19
  * [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-GGML).
20
  * [float16 HF format model for GPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-HF).
21
-
22
  ## How to easily download and use this model in text-generation-webui
23
 
24
  Open the text-generation-webui UI as normal.
@@ -53,7 +64,18 @@ It was created without the `--act-order` parameter. It may have slightly lower i
53
  ```
54
  CUDA_VISIBLE_DEVICES=0 python3 llama.py wizard-vicuna-13B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors wizard-vicuna-13B-GPTQ-4bit.compat.no-act-order.safetensors
55
  ```
56
-
 
 
 
 
 
 
 
 
 
 
 
57
  # Original WizardVicuna-13B model card
58
 
59
  Github page: https://github.com/melodysdreamj/WizardVicunaLM
@@ -68,7 +90,7 @@ I am a big fan of the ideas behind WizardLM and VicunaLM. I particularly like th
68
  ![](https://user-images.githubusercontent.com/21379657/236088663-3fa212c9-0112-4d44-9b01-f16ea093cb67.png)
69
 
70
 
71
- ### Detail
72
 
73
  The questions presented here are not from rigorous tests, but rather, I asked a few questions and requested GPT-4 to score them. The models compared were ChatGPT 3.5, WizardVicunaLM, VicunaLM, and WizardLM, in that order.
74
 
 
6
  - llama
7
  inference: false
8
  ---
9
+ <div style="width: 100%;">
10
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
11
+ </div>
12
+ <div style="display: flex; justify-content: space-between; width: 100%;">
13
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
14
+ <p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
15
+ </div>
16
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
17
+ <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
18
+ </div>
19
+ </div>
20
 
21
  # Wizard-Vicuna-13B-GPTQ
22
 
 
29
  * [4bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-GPTQ).
30
  * [4bit and 5bit GGML models for CPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-GGML).
31
  * [float16 HF format model for GPU inference](https://huggingface.co/TheBloke/wizard-vicuna-13B-HF).
32
+
33
  ## How to easily download and use this model in text-generation-webui
34
 
35
  Open the text-generation-webui UI as normal.
 
64
  ```
65
  CUDA_VISIBLE_DEVICES=0 python3 llama.py wizard-vicuna-13B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors wizard-vicuna-13B-GPTQ-4bit.compat.no-act-order.safetensors
66
  ```
67
+
68
+ ## Want to support my work?
69
+
70
+ I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
71
+
72
+ So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
73
+
74
+ Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
75
+
76
+ * Patreon: coming soon! (just awaiting approval)
77
+ * Ko-Fi: https://ko-fi.com/TheBlokeAI
78
+ * Discord: https://discord.gg/UBgz4VXf
79
  # Original WizardVicuna-13B model card
80
 
81
  Github page: https://github.com/melodysdreamj/WizardVicunaLM
 
90
  ![](https://user-images.githubusercontent.com/21379657/236088663-3fa212c9-0112-4d44-9b01-f16ea093cb67.png)
91
 
92
 
93
+ ### Detail
94
 
95
  The questions presented here are not from rigorous tests, but rather, I asked a few questions and requested GPT-4 to score them. The models compared were ChatGPT 3.5, WizardVicunaLM, VicunaLM, and WizardLM, in that order.
96