anakin87 commited on
Commit
80479ed
1 Parent(s): f87fd4a

change config. json (https://www.reddit.com/r/LocalLLaMA/comments/1cb3q0i/issue_with_with_llama_3_exl2_quant_either_ending/)

Browse files
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -6,7 +6,7 @@
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
- "eos_token_id": 128001,
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
 
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
+ "eos_token_id": 128009,
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,