File size: 9,125 Bytes
b561d2a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6ba3f19
 
 
 
 
 
 
 
 
 
 
 
 
 
b561d2a
 
84c580b
b561d2a
 
 
 
 
0205e9a
b561d2a
 
 
 
 
c041b7e
0ad96c3
92ab0d0
b299872
92ab0d0
b561d2a
 
 
 
 
1bfa258
9abdffe
574d8b7
fd273e3
16abccc
 
9c87da8
79a0eb0
9c87da8
06f5dbd
18e5364
93d0c92
e740a4e
6c6a6ff
e1155ff
d5826d4
275c7ed
8b5ef66
afe9d9b
7c0b348
b561d2a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4f79769
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b561d2a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
---
base_model: deepseek-ai/DeepSeek-V2-Lite-Chat
inference: false
library_name: gguf
pipeline_tag: text-generation
quantized_by: legraphista
tags:
- quantized
- GGUF
- imatrix
- quantization
- imat
- imatrix
- static
---

# DeepSeek-V2-Lite-Chat-IMat-GGUF
_Llama.cpp imatrix quantization of deepseek-ai/DeepSeek-V2-Lite-Chat_

- [DeepSeek-V2-Lite-Chat-IMat-GGUF](#deepseek-v2-lite-chat-imat-gguf)
    - [Files](#files)
        - [IMatrix](#imatrix)
        - [Common Quants](#common-quants)
        - [All Quants](#all-quants)
    - [Downloading using huggingface-cli](#downloading-using-huggingface-cli)
    - [Inference](#inference)
        - [Simple chat template](#simple-chat-template)
        - [Chat template with system prompt](#chat-template-with-system-prompt)
        - [Llama.cpp](#llama-cpp)
    - [FAQ](#faq)
        - [Why is the IMatrix not applied everywhere?](#why-is-the-imatrix-not-applied-everywhere)
        - [How do I merge a split GGUF?](#how-do-i-merge-a-split-gguf)

Original Model: [deepseek-ai/DeepSeek-V2-Lite-Chat](https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite-Chat)  
Original dtype: `BF16` (`bfloat16`)  
Quantized by: llama.cpp fork [PR 7519](https://github.com/ggerganov/llama.cpp/pull/7519)  
IMatrix dataset: [here](https://gist.githubusercontent.com/legraphista/d6d93f1a254bcfc58e0af3777eaec41e/raw/d380e7002cea4a51c33fffd47db851942754e7cc/imatrix.calibration.medium.raw)  

## Files

### IMatrix
Status: ✅ Available  
Link: [here](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/imatrix.dat) 

### Common Quants
| Filename | Quant type | File Size | Status | Uses IMatrix | Is Split |
| -------- | ---------- | --------- | ------ | ------------ | -------- |
| [DeepSeek-V2-Lite-Chat.Q8_0.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q8_0.gguf) | Q8_0 | 16.70GB | ✅ Available | ⚪ No | 📦 No
| [DeepSeek-V2-Lite-Chat.Q6_K.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q6_K.gguf) | Q6_K | 14.07GB | ✅ Available | ⚪ No | 📦 No
| [DeepSeek-V2-Lite-Chat.Q4_K.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q4_K.gguf) | Q4_K | 10.36GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.Q3_K.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q3_K.gguf) | Q3_K | 8.13GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.Q2_K.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q2_K.gguf) | Q2_K | 6.43GB | ✅ Available | 🟢 Yes | 📦 No


### All Quants
| Filename | Quant type | File Size | Status | Uses IMatrix | Is Split |
| -------- | ---------- | --------- | ------ | ------------ | -------- |
| [DeepSeek-V2-Lite-Chat.FP16.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.FP16.gguf) | F16 | 31.42GB | ✅ Available | ⚪ No | 📦 No
| [DeepSeek-V2-Lite-Chat.BF16.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.BF16.gguf) | BF16 | 31.42GB | ✅ Available | ⚪ No | 📦 No
| [DeepSeek-V2-Lite-Chat.Q5_K.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q5_K.gguf) | Q5_K | 11.85GB | ✅ Available | ⚪ No | 📦 No
| [DeepSeek-V2-Lite-Chat.Q5_K_S.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q5_K_S.gguf) | Q5_K_S | 11.14GB | ✅ Available | ⚪ No | 📦 No
| [DeepSeek-V2-Lite-Chat.Q4_K_S.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q4_K_S.gguf) | Q4_K_S | 9.53GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.Q3_K_L.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q3_K_L.gguf) | Q3_K_L | 8.46GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.Q3_K_S.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q3_K_S.gguf) | Q3_K_S | 7.49GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.Q2_K_S.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.Q2_K_S.gguf) | Q2_K_S | 6.46GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ4_NL.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ4_NL.gguf) | IQ4_NL | 8.91GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ4_XS.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ4_XS.gguf) | IQ4_XS | 8.57GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ3_M.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ3_M.gguf) | IQ3_M | 7.55GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ3_S.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ3_S.gguf) | IQ3_S | 7.49GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ3_XS.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ3_XS.gguf) | IQ3_XS | 7.12GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ3_XXS.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ3_XXS.gguf) | IQ3_XXS | 6.96GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ2_M.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ2_M.gguf) | IQ2_M | 6.33GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ2_S.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ2_S.gguf) | IQ2_S | 6.01GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ2_XS.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ2_XS.gguf) | IQ2_XS | 5.97GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ2_XXS.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ2_XXS.gguf) | IQ2_XXS | 5.64GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ1_M.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ1_M.gguf) | IQ1_M | 5.24GB | ✅ Available | 🟢 Yes | 📦 No
| [DeepSeek-V2-Lite-Chat.IQ1_S.gguf](https://huggingface.co/legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF/blob/main/DeepSeek-V2-Lite-Chat.IQ1_S.gguf) | IQ1_S | 4.99GB | ✅ Available | 🟢 Yes | 📦 No


## Downloading using huggingface-cli
First, make sure you have hugginface-cli installed:
```
pip install -U "huggingface_hub[cli]"
```
Then, you can target the specific file you want:
```
huggingface-cli download legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF --include "DeepSeek-V2-Lite-Chat.Q8_0.gguf" --local-dir ./
```
If the model is bigger than 50GB, it will have been split into multiple files. In order to download them all to a local folder, run:
```
huggingface-cli download legraphista/DeepSeek-V2-Lite-Chat-IMat-GGUF --include "DeepSeek-V2-Lite-Chat.Q8_0/*" --local-dir DeepSeek-V2-Lite-Chat.Q8_0
# see FAQ for merging GGUF's
```

---

## Inference

### Simple chat template
```
<|begin▁of▁sentence|>User: {user_message_1}

Assistant: {assistant_message_1}<|end▁of▁sentence|>User: {user_message_2}

Assistant:
```

### Chat template with system prompt
```
<|begin▁of▁sentence|>{system_message}

User: {user_message_1}

Assistant: {assistant_message_1}<|end▁of▁sentence|>User: {user_message_2}

Assistant:
```

### Llama.cpp
```
llama.cpp/main -m DeepSeek-V2-Lite-Chat.Q8_0.gguf --color -i -p "prompt here (according to the chat template)"
```

---

## FAQ

### Why is the IMatrix not applied everywhere?
According to [this investigation](https://www.reddit.com/r/LocalLLaMA/comments/1993iro/ggufs_quants_can_punch_above_their_weights_now/), it appears that lower quantizations are the only ones that benefit from the imatrix input (as per hellaswag results). 

### How do I merge a split GGUF?
1. Make sure you have `gguf-split` available
    - To get hold of `gguf-split`, navigate to https://github.com/ggerganov/llama.cpp/releases
    - Download the appropriate zip for your system from the latest release
    - Unzip the archive and you should be able to find `gguf-split`
2. Locate your GGUF chunks folder (ex: `DeepSeek-V2-Lite-Chat.Q8_0`)
3. Run `gguf-split --merge DeepSeek-V2-Lite-Chat.Q8_0/DeepSeek-V2-Lite-Chat.Q8_0-00001-of-XXXXX.gguf DeepSeek-V2-Lite-Chat.Q8_0.gguf`
    - Make sure to point `gguf-split` to the first chunk of the split.

---

Got a suggestion? Ping me [@legraphista](https://x.com/legraphista)!