actually, it wasn't just Mixtral. Something got broken in the older Llama and Alpaca encoders as well.
am a big fan of Fimbulvetr 10.7B v1.0, and when offloading, my speeds went from 10-13 T/s to a whopping 3 T/s.
It seems to have possibly been fixed in newer versions (am using KCPP as my backened, and they haven't done a build using the newest LCPP code, but BackyardAI has, and I'm now getting 10 T/s).
Sadly, SAO10K took down the un-quanted Fimbul 1.0 repo, so I'm not sure anyone would be able to re-quant and test if that does the trick.