spedrox-sac commited on
Commit
60c2bf3
1 Parent(s): e214489

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -0
README.md ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - fka/awesome-chatgpt-prompts
5
+ language:
6
+ - en
7
+ base_model:
8
+ - Qwen/Qwen2.5-1.5B-Instruct
9
+ pipeline_tag: text-generation
10
+ ---
11
+ # Quantized Qwen2.5-1.5B-Instruct
12
+
13
+ This repository contains 8-bit and 4-bit quantized versions of the Qwen2.5-1.5B-Instruct model using GPTQ. Quantization significantly reduces the model's size and memory footprint, enabling faster inference on resource-constrained devices while maintaining reasonable performance.
14
+
15
+
16
+ ## Model Description
17
+
18
+ The Qwen2.5-1.5B-Instruct is a powerful language model developed by Qwen for instructional tasks. These quantized versions offer a more efficient way to deploy and utilize this model.
19
+
20
+
21
+ ## Quantization Details
22
+
23
+ * **Quantization Method:** GPTQ (Generative Pretrained Transformer Quantization)
24
+ * **Quantization Bits:** 8-bit and 4-bit versions available.
25
+ * **Dataset:** The model was quantized using a subset of the "fka/awesome-chatgpt-prompts" dataset.
26
+
27
+
28
+ ## Usage
29
+
30
+ To use the quantized models, follow these steps:
31
+
32
+ **Install Dependencies:**
33
+ ```bash
34
+ pip install transformers accelerate bitsandbytes auto-gptq optimum
35
+ ```
36
+ ## Performance
37
+
38
+ The quantized models offer a significant reduction in size and memory usage compared to the original model. While there might be a slight decrease in performance, the trade-off is often beneficial for deployment on devices with limited resources.
39
+
40
+
41
+ ## Disclaimer
42
+
43
+ These quantized models are provided for research and experimentation purposes. We do not guarantee their performance or suitability for specific applications.
44
+
45
+
46
+ ## Acknowledgements
47
+
48
+ * **Qwen:** For developing the original Qwen2.5-1.5B-Instruct model.
49
+ * **Hugging Face:** For providing the platform and tools for model sharing and quantization.
50
+ * **GPTQ Authors:** For developing the GPTQ quantization method.