TheBloke commited on
Commit
37c80c6
1 Parent(s): 7a287ea

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -3
README.md CHANGED
@@ -1,14 +1,26 @@
1
  ---
 
2
  inference: false
3
  language:
4
  - en
5
  library_name: transformers
6
  license: llama2
7
  model_creator: The Royal Lab
8
- model_link: https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT
9
  model_name: Pygmalion 2 13B SuperCOT
10
  model_type: llama
11
  pipeline_tag: text-generation
 
 
 
 
 
 
 
 
 
 
 
 
12
  quantized_by: TheBloke
13
  tags:
14
  - llama
@@ -47,6 +59,7 @@ Multiple GPTQ parameter permutations are provided; see Provided Files below for
47
  <!-- repositories-available start -->
48
  ## Repositories available
49
 
 
50
  * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT-GPTQ)
51
  * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT-GGUF)
52
  * [The Royal Lab's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT)
@@ -67,6 +80,7 @@ Below is an instruction that describes a task. Write a response that appropriate
67
 
68
  <!-- prompt-template end -->
69
 
 
70
  <!-- README_GPTQ.md-provided-files start -->
71
  ## Provided files and GPTQ parameters
72
 
@@ -245,7 +259,7 @@ Donaters will get priority support on any and all AI/LLM/model questions and req
245
 
246
  **Special thanks to**: Aemon Algiz.
247
 
248
- **Patreon special mentions**: Russ Johnson, J, alfie_i, Alex, NimbleBox.ai, Chadd, Mandus, Nikolai Manek, Ken Nordquist, ya boyyy, Illia Dulskyi, Viktor Bowallius, vamX, Iucharbius, zynix, Magnesian, Clay Pascal, Pierre Kircher, Enrico Ros, Tony Hughes, Elle, Andrey, knownsqashed, Deep Realms, Jerry Meng, Lone Striker, Derek Yates, Pyrater, Mesiah Bishop, James Bentley, Femi Adebogun, Brandon Frisco, SuperWojo, Alps Aficionado, Michael Dempsey, Vitor Caleffi, Will Dee, Edmond Seymore, usrbinkat, LangChain4j, Kacper Wikieł, Luke Pendergrass, John Detwiler, theTransient, Nathan LeClaire, Tiffany J. Kim, biorpg, Eugene Pentland, Stanislav Ovsiannikov, Fred von Graf, terasurfer, Kalila, Dan Guido, Nitin Borwankar, 阿明, Ai Maven, John Villwock, Gabriel Puliatti, Stephen Murray, Asp the Wyvern, danny, Chris Smitley, ReadyPlayerEmma, S_X, Daniel P. Andersen, Olakabola, Jeffrey Morgan, Imad Khwaja, Caitlyn Gatomon, webtim, Alicia Loh, Trenton Dambrowitz, Swaroop Kallakuri, Erik Bjäreholt, Leonard Tan, Spiking Neurons AB, Luke @flexchar, Ajan Kanaga, Thomas Belote, Deo Leter, RoA, Willem Michiel, transmissions 11, subjectnull, Matthew Berman, Joseph William Delisle, David Ziegler, Michael Davis, Johann-Peter Hartmann, Talal Aujan, senxiiz, Artur Olbinski, Rainer Wilmers, Spencer Kim, Fen Risland, Cap'n Zoog, Rishabh Srivastava, Michael Levine, Geoffrey Montalvo, Sean Connelly, Alexandros Triantafyllidis, Pieter, Gabriel Tamborski, Sam, Subspace Studios, Junyu Yang, Pedro Madruga, Vadim, Cory Kujawski, K, Raven Klaugh, Randy H, Mano Prime, Sebastain Graf, Space Cruiser
249
 
250
 
251
  Thank you to all my generous patrons and donaters!
@@ -263,8 +277,10 @@ This is a merge between:
263
  - [Pygmalion 2 13b](https://huggingface.co/PygmalionAI/pygmalion-2-13b)
264
  - [Ausboss's Llama2 SuperCOT loras](https://huggingface.co/ausboss/llama2-13b-supercot-loras) at a weight of 1.00.
265
 
266
- Quantizations provided by myself:
267
  - [GGUF](https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT-GGUF)
 
 
268
 
269
  The merge was performed by a commandline version of [EzTrainer](https://github.com/CoffeeVampir3/ez-trainer) by CoffeeVampire/Blackroot via [zaraki-tools](https://github.com/CoffeeVampir3/ez-trainer) by Zaraki.
270
 
 
1
  ---
2
+ base_model: https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT
3
  inference: false
4
  language:
5
  - en
6
  library_name: transformers
7
  license: llama2
8
  model_creator: The Royal Lab
 
9
  model_name: Pygmalion 2 13B SuperCOT
10
  model_type: llama
11
  pipeline_tag: text-generation
12
+ prompt_template: 'Below is an instruction that describes a task. Write a response
13
+ that appropriately completes the request.
14
+
15
+
16
+ ### Instruction:
17
+
18
+ {prompt}
19
+
20
+
21
+ ### Response:
22
+
23
+ '
24
  quantized_by: TheBloke
25
  tags:
26
  - llama
 
59
  <!-- repositories-available start -->
60
  ## Repositories available
61
 
62
+ * [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT-AWQ)
63
  * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT-GPTQ)
64
  * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT-GGUF)
65
  * [The Royal Lab's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT)
 
80
 
81
  <!-- prompt-template end -->
82
 
83
+
84
  <!-- README_GPTQ.md-provided-files start -->
85
  ## Provided files and GPTQ parameters
86
 
 
259
 
260
  **Special thanks to**: Aemon Algiz.
261
 
262
+ **Patreon special mentions**: Alicia Loh, Stephen Murray, K, Ajan Kanaga, RoA, Magnesian, Deo Leter, Olakabola, Eugene Pentland, zynix, Deep Realms, Raymond Fosdick, Elijah Stavena, Iucharbius, Erik Bjäreholt, Luis Javier Navarrete Lozano, Nicholas, theTransient, John Detwiler, alfie_i, knownsqashed, Mano Prime, Willem Michiel, Enrico Ros, LangChain4j, OG, Michael Dempsey, Pierre Kircher, Pedro Madruga, James Bentley, Thomas Belote, Luke @flexchar, Leonard Tan, Johann-Peter Hartmann, Illia Dulskyi, Fen Risland, Chadd, S_X, Jeff Scroggin, Ken Nordquist, Sean Connelly, Artur Olbinski, Swaroop Kallakuri, Jack West, Ai Maven, David Ziegler, Russ Johnson, transmissions 11, John Villwock, Alps Aficionado, Clay Pascal, Viktor Bowallius, Subspace Studios, Rainer Wilmers, Trenton Dambrowitz, vamX, Michael Levine, 준교 김, Brandon Frisco, Kalila, Trailburnt, Randy H, Talal Aujan, Nathan Dryer, Vadim, 阿明, ReadyPlayerEmma, Tiffany J. Kim, George Stoitzev, Spencer Kim, Jerry Meng, Gabriel Tamborski, Cory Kujawski, Jeffrey Morgan, Spiking Neurons AB, Edmond Seymore, Alexandros Triantafyllidis, Lone Striker, Cap'n Zoog, Nikolai Manek, danny, ya boyyy, Derek Yates, usrbinkat, Mandus, TL, Nathan LeClaire, subjectnull, Imad Khwaja, webtim, Raven Klaugh, Asp the Wyvern, Gabriel Puliatti, Caitlyn Gatomon, Joseph William Delisle, Jonathan Leane, Luke Pendergrass, SuperWojo, Sebastain Graf, Will Dee, Fred von Graf, Andrey, Dan Guido, Daniel P. Andersen, Nitin Borwankar, Elle, Vitor Caleffi, biorpg, jjj, NimbleBox.ai, Pieter, Matthew Berman, terasurfer, Michael Davis, Alex, Stanislav Ovsiannikov
263
 
264
 
265
  Thank you to all my generous patrons and donaters!
 
277
  - [Pygmalion 2 13b](https://huggingface.co/PygmalionAI/pygmalion-2-13b)
278
  - [Ausboss's Llama2 SuperCOT loras](https://huggingface.co/ausboss/llama2-13b-supercot-loras) at a weight of 1.00.
279
 
280
+ Quantizations provided by us and TheBloke:
281
  - [GGUF](https://huggingface.co/royallab/Pygmalion-2-13b-SuperCOT-GGUF)
282
+ - [GGUF (TheBloke)](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT-GGUF)
283
+ - [GPTQ](https://huggingface.co/TheBloke/Pygmalion-2-13B-SuperCOT-GPTQ)
284
 
285
  The merge was performed by a commandline version of [EzTrainer](https://github.com/CoffeeVampir3/ez-trainer) by CoffeeVampire/Blackroot via [zaraki-tools](https://github.com/CoffeeVampir3/ez-trainer) by Zaraki.
286