Question regarding BOS-token when chatting with the model.

#14
by Ghaandy - opened

Hello!

I came across this model and I must say, I'm very impressed with the results.
While examining the tokenizer_config.json file, I noticed that it specifies < s > as the BOS-token and add_bos_token:true.

Although the model card mentions using the ChatML format, this raised some questions that I hope to receive assistance with.
How should < s > be specified in the prompt/prompts when chatting with the model?

Example 1 - Not at all.
"<|im_start|>User
Question 1?<|im_end|>
<|im_start|>Assistant
Answer 1.<|im_end|>
<|im_start|>User
Question 2 based on question 1?<|im_end|>
<|im_start|>Assistant"

Example 2 - Before each message/prompt.
"< s > <|im_start|>User
I'm human<|im_end|>
< s > <|im_start|>Assistant
I'm Assistant<|im_end|>
< s > |im_start|>User
I'm human<|im_end|>
< s >|im_start|>Assistant
"
Example 3 - At the start of the sequence.
"< s > <|im_start|>User
Question 1?<|im_end|>
<|im_start|>Assistant
Answer 1<|im_end|>
<|im_start|>User
Question 2 based on question 1?<|im_end|
<|im_start|>Assistant"

Thanks.

edit. I can't write the proper BOS-token for some reason without formatting the whole text on here. I've added spaces to make it appear in the text.

Hi there, I am also a bit confused about this and still trying to clear out the confusion myself!

However I think I can answer your question "How should <s> be specified in the prompt/prompts when chatting with the model?".

The answer is: you should not add it yourself. The tokenizer will take care of doing that.

Sign up or log in to comment