TheBloke commited on
Commit
ec8e6a5
1 Parent(s): cd914b3

GPTQ model commit

Browse files
Files changed (1) hide show
  1. config.json +21 -1
config.json CHANGED
@@ -14,6 +14,26 @@
14
  "num_attention_heads": 32,
15
  "num_hidden_layers": 32,
16
  "num_key_value_heads": 8,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  "rms_norm_eps": 1e-05,
18
  "rope_theta": 10000.0,
19
  "sliding_window": 4096,
@@ -24,4 +44,4 @@
24
  "vocab_size": 32000,
25
  "pretraining_tp": 1,
26
  "pad_token_id": 0
27
- }
 
14
  "num_attention_heads": 32,
15
  "num_hidden_layers": 32,
16
  "num_key_value_heads": 8,
17
+ "quantization_config": {
18
+ "batch_size": 1,
19
+ "bits": 4,
20
+ "block_name_to_quantize": "model.layers",
21
+ "damp_percent": 0.1,
22
+ "desc_act": true,
23
+ "disable_exllama": false,
24
+ "group_size": 128,
25
+ "max_input_length": null,
26
+ "model_seqlen": 4096,
27
+ "module_name_preceding_first_block": [
28
+ "model.embed_tokens"
29
+ ],
30
+ "pad_token_id": null,
31
+ "quant_method": "gptq",
32
+ "sym": true,
33
+ "tokenizer": null,
34
+ "true_sequential": true,
35
+ "use_cuda_fp16": true
36
+ },
37
  "rms_norm_eps": 1e-05,
38
  "rope_theta": 10000.0,
39
  "sliding_window": 4096,
 
44
  "vocab_size": 32000,
45
  "pretraining_tp": 1,
46
  "pad_token_id": 0
47
+ }