Update README.md
Browse files
README.md
CHANGED
@@ -12,7 +12,7 @@ This repository contains CPU-optimized GGUF quantizations of the Meta-Llama-3.1-
|
|
12 |
1. Q4_0_4_8 (CPU FMA-Optimized): ~246 GB
|
13 |
2. BF16: ~811 GB
|
14 |
3. Q8_0: ~406 GB
|
15 |
-
4. Q2-
|
16 |
|
17 |
## Use Aria2 for parallelized downloads, links will download 9x faster
|
18 |
|
|
|
12 |
1. Q4_0_4_8 (CPU FMA-Optimized): ~246 GB
|
13 |
2. BF16: ~811 GB
|
14 |
3. Q8_0: ~406 GB
|
15 |
+
4. Q2-Q8 (custom quant I wrote) ~ 165 GB
|
16 |
|
17 |
## Use Aria2 for parallelized downloads, links will download 9x faster
|
18 |
|