TheBloke commited on
Commit
d6c5166
1 Parent(s): b2b8f41

Update config.json for upcoming Transformers support

Browse files
Files changed (1) hide show
  1. config.json +12 -1
config.json CHANGED
@@ -31,6 +31,17 @@
31
  "vocab_size": 32000,
32
  "quantization_config": {
33
  "bits": 8,
 
 
 
 
 
 
 
 
 
 
 
34
  "group_size": -1,
35
  "damp_percent": 0.1,
36
  "desc_act": true,
@@ -40,4 +51,4 @@
40
  "model_file_base_name": "model",
41
  "quant_method": "gptq"
42
  }
43
- }
 
31
  "vocab_size": 32000,
32
  "quantization_config": {
33
  "bits": 8,
34
+ "inside_layer_modules" : [
35
+ ["self_attn.k_proj", "self_attn.v_proj", "self_attn.q_proj"],
36
+ ["self_attn.o_proj"],
37
+ ["block_sparse_moe.experts.0.w1", "block_sparse_moe.experts.0.w2", "block_sparse_moe.experts.0.w3"],
38
+ ["block_sparse_moe.experts.1.w1", "block_sparse_moe.experts.1.w2", "block_sparse_moe.experts.1.w3"],
39
+ ["block_sparse_moe.experts.2.w1", "block_sparse_moe.experts.2.w2", "block_sparse_moe.experts.2.w3"],
40
+ ["block_sparse_moe.experts.3.w1", "block_sparse_moe.experts.3.w2", "block_sparse_moe.experts.3.w3"],
41
+ ["block_sparse_moe.experts.4.w1", "block_sparse_moe.experts.4.w2", "block_sparse_moe.experts.4.w3"],
42
+ ["block_sparse_moe.experts.5.w1", "block_sparse_moe.experts.5.w2", "block_sparse_moe.experts.5.w3"],
43
+ ["block_sparse_moe.experts.6.w1", "block_sparse_moe.experts.6.w2", "block_sparse_moe.experts.6.w3"],
44
+ ["block_sparse_moe.experts.7.w1", "block_sparse_moe.experts.7.w2", "block_sparse_moe.experts.7.w3"]],
45
  "group_size": -1,
46
  "damp_percent": 0.1,
47
  "desc_act": true,
 
51
  "model_file_base_name": "model",
52
  "quant_method": "gptq"
53
  }
54
+ }