uukuguy commited on
Commit
4109ef2
1 Parent(s): 501b1ee

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -10,6 +10,10 @@ license: apache-2.0
10
 
11
  <p><h1> speechless-mistral-moloras-7b </h1></p>
12
 
 
 
 
 
13
  [4-bit GGUF models for CPU+GPU inference](https://huggingface.co/uukuguy/speechless-mistral-moloras-7b/tree/main/GGUF)
14
 
15
  This model is the static version of moloras (Mixture-of-multi-LoRAs) based on the following 6 Mistral-based LoRa modules.
 
10
 
11
  <p><h1> speechless-mistral-moloras-7b </h1></p>
12
 
13
+ * [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/speechless-mistral-moloras-7B-AWQ)
14
+ * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/speechless-mistral-moloras-7B-GPTQ)
15
+ * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/speechless-mistral-moloras-7B-GGUF)
16
+
17
  [4-bit GGUF models for CPU+GPU inference](https://huggingface.co/uukuguy/speechless-mistral-moloras-7b/tree/main/GGUF)
18
 
19
  This model is the static version of moloras (Mixture-of-multi-LoRAs) based on the following 6 Mistral-based LoRa modules.