huseink commited on
Commit
ace9da3
1 Parent(s): 4210309

Update config.json

Browse files

Update the eos_token_id to 128009, replacing the previous value, to address output generation issues with Llama3 instruct, in accordance with updates from the official Meta LLama3 Instruct repository (see https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct/discussions/49).

From <|end_of_text|> to <|eot_id|>

Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -6,7 +6,7 @@
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
- "eos_token_id": 128001,
10
  "hidden_act": "silu",
11
  "hidden_size": 8192,
12
  "initializer_range": 0.02,
 
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
+ "eos_token_id": 128009,
10
  "hidden_act": "silu",
11
  "hidden_size": 8192,
12
  "initializer_range": 0.02,