mike-ravkine commited on
Commit
214ede7
1 Parent(s): 9d2f75e

Fix eos_token_id to align with vocabulary of this model

Browse files

Same problem as the 40b model in this repo.

See https://huggingface.co/TheBloke/falcon-40b-instruct-GPTQ/discussions/8

Upstream discussion: https://huggingface.co/tiiuae/falcon-40b-instruct/discussions/11

Files changed (1) hide show
  1. configuration_RW.py +1 -1
configuration_RW.py CHANGED
@@ -38,7 +38,7 @@ class RWConfig(PretrainedConfig):
38
  initializer_range=0.02,
39
  use_cache=True,
40
  bos_token_id=1,
41
- eos_token_id=2,
42
  apply_residual_connection_post_layernorm=False,
43
  hidden_dropout=0.0,
44
  attention_dropout=0.0,
 
38
  initializer_range=0.02,
39
  use_cache=True,
40
  bos_token_id=1,
41
+ eos_token_id=11,
42
  apply_residual_connection_post_layernorm=False,
43
  hidden_dropout=0.0,
44
  attention_dropout=0.0,