Update README.md
Browse files
README.md
CHANGED
@@ -26,7 +26,11 @@ You need [autoGPTQ](https://github.com/PanQiWei/AutoGPTQ) library to use this mo
|
|
26 |
|
27 |
## Other Quantized Model
|
28 |
|
29 |
-
|
|
|
|
|
|
|
|
|
30 |
If you want to run it in a CPU-only environment, you may want to check this.
|
31 |
|
32 |
(1)[mmnga's gguf version](https://huggingface.co/mmnga/ELYZA-japanese-Llama-2-7b-fast-instruct-gguf)
|
|
|
26 |
|
27 |
## Other Quantized Model
|
28 |
|
29 |
+
### New!
|
30 |
+
[dahara1/ELYZA-japanese-Llama-2-7b-instruct-AWQ](https://huggingface.co/dahara1/ELYZA-japanese-Llama-2-7b-instruct-AWQ) is newly published.
|
31 |
+
The awq model has improved ability to follow instructions, so please try it.
|
32 |
+
|
33 |
+
There are another two [llama.cpp](https://github.com/ggerganov/llama.cpp) version quantized model.
|
34 |
If you want to run it in a CPU-only environment, you may want to check this.
|
35 |
|
36 |
(1)[mmnga's gguf version](https://huggingface.co/mmnga/ELYZA-japanese-Llama-2-7b-fast-instruct-gguf)
|