GGML request
#1
by
DaigoNorr
- opened
Hey, seems like the bigger context is available for llama.cpp and koboldcpp. Could you quantize it in GGML? Thanks!
Still need the ggml versions, should be possible now
Hey, seems like the bigger context is available for llama.cpp and koboldcpp. Could you quantize it in GGML? Thanks!
Still need the ggml versions, should be possible now