beomi commited on
Commit
1e21ef2
1 Parent(s): c518e87
Files changed (1) hide show
  1. config.json +2 -2
config.json CHANGED
@@ -6,7 +6,7 @@
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 1,
9
- "eos_token_id": 2,
10
  "head_dim": 80,
11
  "hidden_act": "silu",
12
  "hidden_size": 2560,
@@ -22,7 +22,7 @@
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
25
- "factor": 32.0,
26
  "high_freq_factor": 4.0,
27
  "low_freq_factor": 1.0,
28
  "original_max_position_embeddings": 8192,
 
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 1,
9
+ "eos_token_id": 361,
10
  "head_dim": 80,
11
  "hidden_act": "silu",
12
  "hidden_size": 2560,
 
22
  "pretraining_tp": 1,
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
25
+ "factor": 8.0,
26
  "high_freq_factor": 4.0,
27
  "low_freq_factor": 1.0,
28
  "original_max_position_embeddings": 8192,