wrong padding token

#2
by aliencaocao - opened

per the author's training code, unk is also used as padding token. However, this repo uses a new , untrained one.

Llava Hugging Face org

Thanks for reporting, will fix in https://github.com/huggingface/transformers/pull/29586

Llava Hugging Face org

Not sure using <unk> is a good practice, but whatever the content if the model outputs 0 for the pad_token_idx does it make a difference? Appart from unkown tokens being able to pop randomly in a text prompt, which does not sound like padding no?

Sign up or log in to comment