Update README.md
Browse files
README.md
CHANGED
@@ -14,9 +14,7 @@ license: cc-by-nc-4.0
|
|
14 |
model_creator: Charles Goddard
|
15 |
model_name: MixtralRPChat ZLoss
|
16 |
model_type: mixtral
|
17 |
-
prompt_template: '***System: {system_message} ***Query: {prompt} ***Response:
|
18 |
-
|
19 |
-
'
|
20 |
quantized_by: TheBloke
|
21 |
tags:
|
22 |
- mixtral
|
@@ -66,7 +64,7 @@ AWQ is an efficient, accurate and blazing-fast low-bit weight quantization metho
|
|
66 |
|
67 |
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
|
68 |
|
69 |
-
AWQ models are supported by (note that
|
70 |
|
71 |
- [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
|
72 |
- [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
|
|
|
14 |
model_creator: Charles Goddard
|
15 |
model_name: MixtralRPChat ZLoss
|
16 |
model_type: mixtral
|
17 |
+
prompt_template: '***System: {system_message} ***Query: {prompt} ***Response:'
|
|
|
|
|
18 |
quantized_by: TheBloke
|
19 |
tags:
|
20 |
- mixtral
|
|
|
64 |
|
65 |
AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
|
66 |
|
67 |
+
AWQ models are supported by (note that not all of these may support Mixtral models yet):
|
68 |
|
69 |
- [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
|
70 |
- [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
|