Text Generation
Transformers
Safetensors
minimax_m2
conversational
custom_code
fp8
xuebi commited on
Commit
6d30eb0
·
1 Parent(s): c2b7e11

update: prepare for transformers

Browse files

Signed-off-by: xuebi <xuebi@minimaxi.com>

Files changed (2) hide show
  1. config.json +6 -1
  2. generation_config.json +2 -0
config.json CHANGED
@@ -79,7 +79,7 @@
79
  "layernorm_mlp_beta": 1.0,
80
  "max_position_embeddings": 196608,
81
  "mlp_intermediate_size": 8192,
82
- "model_type": "minimax",
83
  "mtp_transformer_layers": 1,
84
  "num_attention_heads": 48,
85
  "num_experts_per_tok": 8,
@@ -96,6 +96,11 @@
96
  "weight_block_size": [
97
  128,
98
  128
 
 
 
 
 
99
  ]
100
  },
101
  "rms_norm_eps": 1e-06,
 
79
  "layernorm_mlp_beta": 1.0,
80
  "max_position_embeddings": 196608,
81
  "mlp_intermediate_size": 8192,
82
+ "model_type": "minimax_m2",
83
  "mtp_transformer_layers": 1,
84
  "num_attention_heads": 48,
85
  "num_experts_per_tok": 8,
 
96
  "weight_block_size": [
97
  128,
98
  128
99
+ ],
100
+ "modules_to_not_convert": [
101
+ "gate",
102
+ "e_score_correction_bias",
103
+ "lm_head"
104
  ]
105
  },
106
  "rms_norm_eps": 1e-06,
generation_config.json CHANGED
@@ -1,5 +1,7 @@
1
  {
 
2
  "do_sample": true,
 
3
  "temperature": 1.0,
4
  "top_p": 0.95,
5
  "top_k": 40,
 
1
  {
2
+ "bos_token_id": 200019,
3
  "do_sample": true,
4
+ "eos_token_id": 200020,
5
  "temperature": 1.0,
6
  "top_p": 0.95,
7
  "top_k": 40,