Commit
β’
f89f989
1
Parent(s):
533d142
Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -28,7 +28,7 @@ _Llama.cpp imatrix quantization of IEITYuan/Yuan2-M32-hf_
|
|
28 |
|
29 |
Original Model: [IEITYuan/Yuan2-M32-hf](https://huggingface.co/IEITYuan/Yuan2-M32-hf)
|
30 |
Original dtype: `BF16` (`bfloat16`)
|
31 |
-
Quantized by: llama
|
32 |
IMatrix dataset: [here](https://gist.githubusercontent.com/legraphista/d6d93f1a254bcfc58e0af3777eaec41e/raw/d380e7002cea4a51c33fffd47db851942754e7cc/imatrix.calibration.medium.raw)
|
33 |
|
34 |
- [Files](#files)
|
@@ -63,8 +63,7 @@ Link: [here](https://huggingface.co/legraphista/Yuan2-M32-hf-IMat-GGUF/blob/main
|
|
63 |
### All Quants
|
64 |
| Filename | Quant type | File Size | Status | Uses IMatrix | Is Split |
|
65 |
| -------- | ---------- | --------- | ------ | ------------ | -------- |
|
66 |
-
| Yuan2-M32-hf.
|
67 |
-
| Yuan2-M32-hf.FP16 | F16 | - | β³ Processing | βͺ Static | -
|
68 |
| [Yuan2-M32-hf.Q8_0.gguf](https://huggingface.co/legraphista/Yuan2-M32-hf-IMat-GGUF/blob/main/Yuan2-M32-hf.Q8_0.gguf) | Q8_0 | 42.93GB | β
Available | βͺ Static | π¦ No
|
69 |
| [Yuan2-M32-hf.Q6_K.gguf](https://huggingface.co/legraphista/Yuan2-M32-hf-IMat-GGUF/blob/main/Yuan2-M32-hf.Q6_K.gguf) | Q6_K | 33.23GB | β
Available | βͺ Static | π¦ No
|
70 |
| Yuan2-M32-hf.Q5_K | Q5_K | - | β³ Processing | βͺ Static | -
|
|
|
28 |
|
29 |
Original Model: [IEITYuan/Yuan2-M32-hf](https://huggingface.co/IEITYuan/Yuan2-M32-hf)
|
30 |
Original dtype: `BF16` (`bfloat16`)
|
31 |
+
Quantized by: [https://github.com/IEIT-Yuan/3rd_party/tree/main/llama-cpp](https://github.com/IEIT-Yuan/3rd_party/tree/main/llama-cpp)
|
32 |
IMatrix dataset: [here](https://gist.githubusercontent.com/legraphista/d6d93f1a254bcfc58e0af3777eaec41e/raw/d380e7002cea4a51c33fffd47db851942754e7cc/imatrix.calibration.medium.raw)
|
33 |
|
34 |
- [Files](#files)
|
|
|
63 |
### All Quants
|
64 |
| Filename | Quant type | File Size | Status | Uses IMatrix | Is Split |
|
65 |
| -------- | ---------- | --------- | ------ | ------------ | -------- |
|
66 |
+
| [Yuan2-M32-hf.FP16/*](https://huggingface.co/legraphista/Yuan2-M32-hf-IMat-GGUF/tree/main/Yuan2-M32-hf.FP16) | F16 | 80.12GB | β
Available | βͺ Static | β Yes
|
|
|
67 |
| [Yuan2-M32-hf.Q8_0.gguf](https://huggingface.co/legraphista/Yuan2-M32-hf-IMat-GGUF/blob/main/Yuan2-M32-hf.Q8_0.gguf) | Q8_0 | 42.93GB | β
Available | βͺ Static | π¦ No
|
68 |
| [Yuan2-M32-hf.Q6_K.gguf](https://huggingface.co/legraphista/Yuan2-M32-hf-IMat-GGUF/blob/main/Yuan2-M32-hf.Q6_K.gguf) | Q6_K | 33.23GB | β
Available | βͺ Static | π¦ No
|
69 |
| Yuan2-M32-hf.Q5_K | Q5_K | - | β³ Processing | βͺ Static | -
|