Run with full 128k context in 24G vram

#3
by meigami - opened

its a amazing model but need too much vram, I cannot run with full 128k context in 24G vram
Do you have any plans to add lighter versions? @TheBloke @LoneStriker

Sign up or log in to comment