File size: 1,352 Bytes
ea9d312
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
**SqueezeLLM** is a post-training quantization framework that incorporates a new method called Dense-and-Sparse Quantization to enable efficient LLM serving.

**TLDR:** Deploying LLMs is difficult due to their large memory size. This can be addressed with reduced precision quantization. 
But a naive method hurts performance. We address this with a new Dense-and-Sparse Quantization method. 
Dense-and-Sparse splits weight matrices into two components: A dense component that can be heavily quantized without affecting model performance, 
as well as a sparse part that preserves sensitive and outlier parts of the weight matrices With this approach, 
we are able to serve larger models with smaller memory footprint, the same latency, and yet higher accuracy and quality. 
For more details please check out our [paper](https://arxiv.org/pdf/2306.07629.pdf).


## Model description

4-bit quantized OPT 13B model using SqueezeLLM. More details can be found in the [paper](https://arxiv.org/pdf/2306.07629.pdf). 

* **Base Model:** [OPT 13B](https://arxiv.org/abs/2205.01068)
* **Bitwidth:** 4-bit
* **Sparsity Level:** 0.5%

## Links

* **Paper**: [https://arxiv.org/pdf/2306.07629.pdf](https://arxiv.org/pdf/2306.07629.pdf)
* **Code**: [https://github.com/SqueezeAILab/SqueezeLLM](https://github.com/SqueezeAILab/SqueezeLLM) 


---
license: other
---