TheBloke commited on
Commit
12190f7
1 Parent(s): 2580c95

Update for Transformers GPTQ support

Browse files
README.md CHANGED
@@ -9,17 +9,20 @@ tags:
9
  inference: false
10
  ---
11
  <!-- header start -->
12
- <div style="width: 100%;">
13
- <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
 
14
  </div>
15
  <div style="display: flex; justify-content: space-between; width: 100%;">
16
  <div style="display: flex; flex-direction: column; align-items: flex-start;">
17
- <p><a href="https://discord.gg/Jq4vkcDakD">Chat & support: my new Discord server</a></p>
18
  </div>
19
  <div style="display: flex; flex-direction: column; align-items: flex-end;">
20
- <p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
21
  </div>
22
  </div>
 
 
23
  <!-- header end -->
24
 
25
  # medalpaca-13B GPTQ 4bit
@@ -104,11 +107,12 @@ The above commands assume you have installed all dependencies for GPTQ-for-LLaMa
104
  If you can't update GPTQ-for-LLaMa to the latest Triton branch, or don't want to, you can use `medalpaca-13B-GPTQ-4bit-128g.no-act-order.safetensors` as mentioned above, which should work without any upgrades to text-generation-webui.
105
 
106
  <!-- footer start -->
 
107
  ## Discord
108
 
109
  For further support, and discussions on these models and AI in general, join us at:
110
 
111
- [TheBloke AI's Discord server](https://discord.gg/Jq4vkcDakD)
112
 
113
  ## Thanks, and how to contribute.
114
 
@@ -123,9 +127,15 @@ Donaters will get priority support on any and all AI/LLM/model questions and req
123
  * Patreon: https://patreon.com/TheBlokeAI
124
  * Ko-Fi: https://ko-fi.com/TheBlokeAI
125
 
126
- **Patreon special mentions**: Aemon Algiz, Dmitriy Samsonov, Nathan LeClaire, Trenton Dambrowitz, Mano Prime, David Flickinger, vamX, Nikolai Manek, senxiiz, Khalefa Al-Ahmad, Illia Dulskyi, Jonathan Leane, Talal Aujan, V. Lukas, Joseph William Delisle, Pyrater, Oscar Rangel, Lone Striker, Luke Pendergrass, Eugene Pentland, Sebastain Graf, Johann-Peter Hartman.
 
 
 
127
 
128
  Thank you to all my generous patrons and donaters!
 
 
 
129
  <!-- footer end -->
130
  # Original model card: MedAlpaca 13b
131
 
 
9
  inference: false
10
  ---
11
  <!-- header start -->
12
+ <!-- 200823 -->
13
+ <div style="width: auto; margin-left: auto; margin-right: auto">
14
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
15
  </div>
16
  <div style="display: flex; justify-content: space-between; width: 100%;">
17
  <div style="display: flex; flex-direction: column; align-items: flex-start;">
18
+ <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p>
19
  </div>
20
  <div style="display: flex; flex-direction: column; align-items: flex-end;">
21
+ <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
22
  </div>
23
  </div>
24
+ <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div>
25
+ <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
26
  <!-- header end -->
27
 
28
  # medalpaca-13B GPTQ 4bit
 
107
  If you can't update GPTQ-for-LLaMa to the latest Triton branch, or don't want to, you can use `medalpaca-13B-GPTQ-4bit-128g.no-act-order.safetensors` as mentioned above, which should work without any upgrades to text-generation-webui.
108
 
109
  <!-- footer start -->
110
+ <!-- 200823 -->
111
  ## Discord
112
 
113
  For further support, and discussions on these models and AI in general, join us at:
114
 
115
+ [TheBloke AI's Discord server](https://discord.gg/theblokeai)
116
 
117
  ## Thanks, and how to contribute.
118
 
 
127
  * Patreon: https://patreon.com/TheBlokeAI
128
  * Ko-Fi: https://ko-fi.com/TheBlokeAI
129
 
130
+ **Special thanks to**: Aemon Algiz.
131
+
132
+ **Patreon special mentions**: Sam, theTransient, Jonathan Leane, Steven Wood, webtim, Johann-Peter Hartmann, Geoffrey Montalvo, Gabriel Tamborski, Willem Michiel, John Villwock, Derek Yates, Mesiah Bishop, Eugene Pentland, Pieter, Chadd, Stephen Murray, Daniel P. Andersen, terasurfer, Brandon Frisco, Thomas Belote, Sid, Nathan LeClaire, Magnesian, Alps Aficionado, Stanislav Ovsiannikov, Alex, Joseph William Delisle, Nikolai Manek, Michael Davis, Junyu Yang, K, J, Spencer Kim, Stefan Sabev, Olusegun Samson, transmissions 11, Michael Levine, Cory Kujawski, Rainer Wilmers, zynix, Kalila, Luke @flexchar, Ajan Kanaga, Mandus, vamX, Ai Maven, Mano Prime, Matthew Berman, subjectnull, Vitor Caleffi, Clay Pascal, biorpg, alfie_i, 阿明, Jeffrey Morgan, ya boyyy, Raymond Fosdick, knownsqashed, Olakabola, Leonard Tan, ReadyPlayerEmma, Enrico Ros, Dave, Talal Aujan, Illia Dulskyi, Sean Connelly, senxiiz, Artur Olbinski, Elle, Raven Klaugh, Fen Risland, Deep Realms, Imad Khwaja, Fred von Graf, Will Dee, usrbinkat, SuperWojo, Alexandros Triantafyllidis, Swaroop Kallakuri, Dan Guido, John Detwiler, Pedro Madruga, Iucharbius, Viktor Bowallius, Asp the Wyvern, Edmond Seymore, Trenton Dambrowitz, Space Cruiser, Spiking Neurons AB, Pyrater, LangChain4j, Tony Hughes, Kacper Wikieł, Rishabh Srivastava, David Ziegler, Luke Pendergrass, Andrey, Gabriel Puliatti, Lone Striker, Sebastain Graf, Pierre Kircher, Randy H, NimbleBox.ai, Vadim, danny, Deo Leter
133
+
134
 
135
  Thank you to all my generous patrons and donaters!
136
+
137
+ And thank you again to a16z for their generous grant.
138
+
139
  <!-- footer end -->
140
  # Original model card: MedAlpaca 13b
141
 
config.json CHANGED
@@ -19,5 +19,15 @@
19
  "torch_dtype": "float32",
20
  "transformers_version": "4.28.0.dev0",
21
  "use_cache": true,
22
- "vocab_size": 32001
23
- }
 
 
 
 
 
 
 
 
 
 
 
19
  "torch_dtype": "float32",
20
  "transformers_version": "4.28.0.dev0",
21
  "use_cache": true,
22
+ "vocab_size": 32001,
23
+ "quantization_config": {
24
+ "bits": 4,
25
+ "group_size": 128,
26
+ "damp_percent": 0.01,
27
+ "desc_act": false,
28
+ "sym": true,
29
+ "true_sequential": true,
30
+ "model_file_base_name": "model",
31
+ "quant_method": "gptq"
32
+ }
33
+ }
medalpaca-13B-GPTQ-4bit-128g.compat.no-act-order.safetensors → model.safetensors RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47dc87220d18238e481e7cca868e53c59ca9e834b24094cde5348485f8f19e98
3
- size 7255179698
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:174ab73c777d5fb47b746d8d6f7a01a0f94e2fe7882e4089acce7d3117f91946
3
+ size 7255179752
quantize_config.json CHANGED
@@ -4,5 +4,6 @@
4
  "damp_percent": 0.01,
5
  "desc_act": false,
6
  "sym": true,
7
- "true_sequential": true
 
8
  }
 
4
  "damp_percent": 0.01,
5
  "desc_act": false,
6
  "sym": true,
7
+ "true_sequential": true,
8
+ "model_file_base_name": "model"
9
  }