exllama requires that pad_token_id be specified in config.json

#2
by mike-ravkine - opened

Hi @TheBloke ,

exllama crashes unless pad_token_id is set in config.json: https://github.com/turboderp/exllama/blob/master/model.py#L52

I have filled in the value of the [PAD] token here as per https://huggingface.co/TheBloke/WizardCoder-Guanaco-15B-V1.1-GPTQ/blob/main/special_tokens_map.json#L25

This issue is likely unique to the GPTQ quant but would affect some of the other branches that have exllama support I think.

Thanks as always,
--Mike

But hang on, this is not a Llama model. So Exllama won't support it anyway, will it?

@TheBloke Ahh you're right they've added Llama2 but for some reason still holding out on BigCode support. This indeed wont help much until then..

mike-ravkine changed pull request status to closed

Sign up or log in to comment