GGML request

#1
by DaigoNorr - opened

Hey, seems like the bigger context is available for llama.cpp and koboldcpp. Could you quantize it in GGML? Thanks!

Still need the ggml versions, should be possible now

Sign up or log in to comment