Context length

#5
by mrfakename - opened

Hi, great model - what’s the context length?

32K i believe. That remains unaffected from zephyr generation, I followed their gym.

Hi, zephyr is 8k right?

i dont have enough GPU to test such thing.. do u ?
But give it a shot, the tok/mod states 32K

fblgit changed discussion status to closed

Nah, I don’t have a GPU, but thanks for the info!

Using Exllamav2_HF and alpha=1, the model breaks down above 8k tokens. When I set alpha=2.5, it worked with a 16k token limit.

I think it would be useful to mention in the model card that the base context is 8k.

hmmmmmm.. Interesting... can u share the full config ? I will update and cite you on the readme, thanks

No need to mention me. I used the exllamav2_HF loader, 8bpw-h8 exl2 quant, simple-1 preset.

Sign up or log in to comment