What pad token should I use for fine tuning?

#10
by faizsameerahmed96 - opened

I am working on fine tuning this model. In the Phi-3-mini, we were recommended to use the unk_token for padding. However it looks like this is no longer present in the tokenizer.

Microsoft org

Hi
You can try using the <|dummy_id_0|> token. Specifically, while creating the tokenizer, you can do something like

tokenizer = AutoTokenizer.from_pretrained("microsoft/Phi-3-small-8k-instruct", trust_remote_code=True, pad_token="<|dummy_id_0|>")
# tokenizer.pad_token should give <|dummy_id_0|>, and tokenizer.pad_token_id should give 100264
bapatra changed discussion status to closed

Sign up or log in to comment