Warning: This model is experimental and unpredictable
CybersurferNyandroidLexicat-8x7B (I was in a silly mood when I made this edition)
Is a linear merge of the following models:
crestf411/daybreak-mixtral-8x7b-v1.0-hf 30%
Experimental unreleased merge 10%
I find its output as an assistant to be less dry and it is stable and imaginative in brief roleplay testing. Tested with simple sampling, requires rerolls but when it's good it's good. I can't say how well it will be when the context fills up but I was pleasantly surprised.
It definitely has one of the most varied lexicons out of any Mixtral Instruct based model I've tested so far with excellent attention to detail with respect to context.
It likes Libra styled prompt formats with [INST] context [/INST] formatting
As always tested in Q8 (not included)
- Downloads last month
- 6
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.