bartowski commited on
Commit
0ef8fdb
1 Parent(s): c20ff15

Llamacpp quants

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Nexus-IKM-Mistral-7B-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Nexus-IKM-Mistral-7B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Nexus-IKM-Mistral-7B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Nexus-IKM-Mistral-7B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Nexus-IKM-Mistral-7B-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Nexus-IKM-Mistral-7B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Nexus-IKM-Mistral-7B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Nexus-IKM-Mistral-7B-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Nexus-IKM-Mistral-7B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Nexus-IKM-Mistral-7B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Nexus-IKM-Mistral-7B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Nexus-IKM-Mistral-7B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
Nexus-IKM-Mistral-7B-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b879fa0efe6f6c46c2871761e09d7032453ac6bc48d4fe88758377ee77b33f53
3
+ size 2719242496
Nexus-IKM-Mistral-7B-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb198c57af4ccbb0037893f91cb60c522d9f9537efeddb01b6a8ffddfd89ce7a
3
+ size 3822024960
Nexus-IKM-Mistral-7B-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a37c40a81323d50ddb68b86c9dca0425a24ca9eadea17156f4ef79766210b22
3
+ size 3518986496
Nexus-IKM-Mistral-7B-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9471e0f4bc6c44b400de2acbb2c06901ee7562e791ea42159e0c0ca5fa111b90
3
+ size 3164567808
Nexus-IKM-Mistral-7B-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9cf2b4489f324c21f32e713d1b34fc00a2831cd3c0e14a28957647ddc15c2d6
3
+ size 4108916992
Nexus-IKM-Mistral-7B-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddeb10bea6fb912de71efbf3d574944b10e8ce5771c41435068670a87a92aa1d
3
+ size 4368439552
Nexus-IKM-Mistral-7B-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce4a3fb5354b8994404e769c86dad38777d1be2ec71e97930961dc9c7596c5e0
3
+ size 4140374272
Nexus-IKM-Mistral-7B-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a39b65666ae0e661d5554afebb1956cb6af96297153ab5fef34b4b3655dd3352
3
+ size 4997716224
Nexus-IKM-Mistral-7B-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06c7afe7274840d9b5b1b7f9ad401ce0ce72a76f109bf28ee78f6670071cd3d1
3
+ size 5131409664
Nexus-IKM-Mistral-7B-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90b9cccdba6785917c48ba1fea24951c5d875f178c736afa5c25c0697b3fcbd1
3
+ size 4997716224
Nexus-IKM-Mistral-7B-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61efd3c1902e98edc8a51a93bcd6e9a340b2d820331b17d8b42ea8b27659fb66
3
+ size 5942065408
Nexus-IKM-Mistral-7B-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a695d5841de0a27c55875ea17b8f2b2e734d942b49091821e2b3339524a7d8cf
3
+ size 7695857920
README.md ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ library_name: transformers
4
+ datasets:
5
+ - Severian/Internal-Knowledge-Map
6
+ pipeline_tag: text-generation
7
+ quantized_by: bartowski
8
+ ---
9
+
10
+ ## Llamacpp Quantizations of Nexus-IKM-Mistral-7B
11
+
12
+ Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b2354">b2354</a> for quantization.
13
+
14
+ Original model: https://huggingface.co/Severian/Nexus-IKM-Mistral-7B
15
+
16
+ Download a file (not the whole branch) from below:
17
+
18
+ | Filename | Quant type | File Size | Description |
19
+ | -------- | ---------- | --------- | ----------- |
20
+ | [Nexus-IKM-Mistral-7B-Q8_0.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q8_0.gguf) | Q8_0 | 7.69GB | Extremely high quality, generally unneeded but max available quant. |
21
+ | [Nexus-IKM-Mistral-7B-Q6_K.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q6_K.gguf) | Q6_K | 5.94GB | Very high quality, near perfect, *recommended*. |
22
+ | [Nexus-IKM-Mistral-7B-Q5_K_M.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q5_K_M.gguf) | Q5_K_M | 5.13GB | High quality, very usable. |
23
+ | [Nexus-IKM-Mistral-7B-Q5_K_S.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q5_K_S.gguf) | Q5_K_S | 4.99GB | High quality, very usable. |
24
+ | [Nexus-IKM-Mistral-7B-Q5_0.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q5_0.gguf) | Q5_0 | 4.99GB | High quality, older format, generally not recommended. |
25
+ | [Nexus-IKM-Mistral-7B-Q4_K_M.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q4_K_M.gguf) | Q4_K_M | 4.36GB | Good quality, similar to 4.25 bpw. |
26
+ | [Nexus-IKM-Mistral-7B-Q4_K_S.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q4_K_S.gguf) | Q4_K_S | 4.14GB | Slightly lower quality with small space savings. |
27
+ | [Nexus-IKM-Mistral-7B-Q4_0.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q4_0.gguf) | Q4_0 | 4.10GB | Decent quality, older format, generally not recommended. |
28
+ | [Nexus-IKM-Mistral-7B-Q3_K_L.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q3_K_L.gguf) | Q3_K_L | 3.82GB | Lower quality but usable, good for low RAM availability. |
29
+ | [Nexus-IKM-Mistral-7B-Q3_K_M.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q3_K_M.gguf) | Q3_K_M | 3.51GB | Even lower quality. |
30
+ | [Nexus-IKM-Mistral-7B-Q3_K_S.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q3_K_S.gguf) | Q3_K_S | 3.16GB | Low quality, not recommended. |
31
+ | [Nexus-IKM-Mistral-7B-Q2_K.gguf](https://huggingface.co/bartowski/Nexus-IKM-Mistral-7B-GGUF//main/Nexus-IKM-Mistral-7B-Q2_K.gguf) | Q2_K | 2.71GB | Extremely low quality, *not* recommended.
32
+
33
+ Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski