shuyuej commited on
Commit
bb01e02
1 Parent(s): 91d41aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -3
README.md CHANGED
@@ -1,3 +1,43 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+
5
+ # The INT-8 Quantized Mistral Nemo Instruct 2407 Model
6
+
7
+ Original Base Model: `mistralai/Mistral-Nemo-Instruct-2407`.<br>
8
+ Link: [https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407](https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407)
9
+
10
+ ## Special Notice
11
+
12
+ The INT-4 Quantized Mistral Nemo Instruct 2407 Model: [https://huggingface.co/shuyuej/Mistral-Nemo-Instruct-2407-GPTQ](https://huggingface.co/shuyuej/Mistral-Nemo-Instruct-2407-GPTQ)
13
+
14
+ ## Quantization Configurations
15
+ ```
16
+ "quantization_config": {
17
+ "batch_size": 1,
18
+ "bits": 8,
19
+ "block_name_to_quantize": null,
20
+ "cache_block_outputs": true,
21
+ "damp_percent": 0.1,
22
+ "dataset": null,
23
+ "desc_act": false,
24
+ "exllama_config": {
25
+ "version": 1
26
+ },
27
+ "group_size": 128,
28
+ "max_input_length": null,
29
+ "model_seqlen": null,
30
+ "module_name_preceding_first_block": null,
31
+ "modules_in_block_to_quantize": null,
32
+ "pad_token_id": null,
33
+ "quant_method": "gptq",
34
+ "sym": true,
35
+ "tokenizer": null,
36
+ "true_sequential": true,
37
+ "use_cuda_fp16": false,
38
+ "use_exllama": true
39
+ },
40
+ ```
41
+
42
+ ## Source Codes
43
+ Source Codes: [https://github.com/vkola-lab/medpodgpt/tree/main/quantization](https://github.com/vkola-lab/medpodgpt/tree/main/quantization).