Text Generation
Transformers
PyTorch
mistral
openchat
C-RLFT
conversational
Inference Endpoints
text-generation-inference
TheBloke commited on
Commit
6055487
1 Parent(s): f85ea3d

EOS should be 32000

Browse files

EOS should be 32000, otherwise generation doesn't terminate on the first `<|endofturn|>` token, making it seem like generation has frozen (in fact it's generating endless `<|endofturn|>` tokens until it reaches the max token limit)

Files changed (1) hide show
  1. tokenizer_config.json +1 -1
tokenizer_config.json CHANGED
@@ -49,7 +49,7 @@
49
  ],
50
  "bos_token": "<s>",
51
  "clean_up_tokenization_spaces": false,
52
- "eos_token": "</s>",
53
  "legacy": true,
54
  "model_max_length": 1000000000000000019884624838656,
55
  "pad_token": null,
 
49
  ],
50
  "bos_token": "<s>",
51
  "clean_up_tokenization_spaces": false,
52
+ "eos_token": "<|end_of_turn|>",
53
  "legacy": true,
54
  "model_max_length": 1000000000000000019884624838656,
55
  "pad_token": null,