mindrage commited on
Commit
992458b
1 Parent(s): 9a903c8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -10,7 +10,11 @@ library_name: transformers
10
  # GGML of:
11
  Manticore-13b-Chat-Pyg by [openaccess-ai-collective](https://huggingface.co/openaccess-ai-collective/manticore-13b-chat-pyg) with the Guanaco 13b qLoRa by [TimDettmers](https://huggingface.co/timdettmers/guanaco-13b) applied through [Monero](https://huggingface.co/Monero/Manticore-13b-Chat-Pyg-Guanaco), quantized by [mindrage](https://huggingface.co/mindrage), uncensored
12
 
13
- (q4_0, q5_0 and q8_0 versions available)
 
 
 
 
14
 
15
  [link to GPTQ Version](https://huggingface.co/mindrage/Manticore-13B-Chat-Pyg-Guanaco-GPTQ-4bit-128g.no-act-order.safetensors)
16
 
 
10
  # GGML of:
11
  Manticore-13b-Chat-Pyg by [openaccess-ai-collective](https://huggingface.co/openaccess-ai-collective/manticore-13b-chat-pyg) with the Guanaco 13b qLoRa by [TimDettmers](https://huggingface.co/timdettmers/guanaco-13b) applied through [Monero](https://huggingface.co/Monero/Manticore-13b-Chat-Pyg-Guanaco), quantized by [mindrage](https://huggingface.co/mindrage), uncensored
12
 
13
+ 12.06.2023: Added versions quantized with the new method (less precision loss relative to compression ratio, but slower (for now)):
14
+ q2_K, q3_KM, q4_KS, q4_KM, q5_KS
15
+
16
+ Old Quant method:
17
+ q4_0, q5_0 and q8_0 versions available
18
 
19
  [link to GPTQ Version](https://huggingface.co/mindrage/Manticore-13B-Chat-Pyg-Guanaco-GPTQ-4bit-128g.no-act-order.safetensors)
20