what about an quantized version so we can load in Exlama with large context size?

#1
by DQ83 - opened

i think coding models should all have large context sizes. A lot of models like losslessmega encoder , CodeUp, StableCode....dont work with Exlama tho :(
any hope for this one?
@TheBloke

Sign up or log in to comment