Edit model card

Warning: This model is experimental and unpredictable

CybersurferNyandroidLexicat-8x7B (I was in a silly mood when I made this edition)

Is a linear merge of the following models:

Verdict-DADA-8x7B 60%

crestf411/daybreak-mixtral-8x7b-v1.0-hf 30%

Experimental unreleased merge 10%

I find its output as an assistant to be less dry and it is stable and imaginative in brief roleplay testing. Tested with simple sampling, requires rerolls but when it's good it's good. I can't say how well it will be when the context fills up but I was pleasantly surprised.

It definitely has one of the most varied lexicons out of any Mixtral Instruct based model I've tested so far with excellent attention to detail with respect to context.

It likes Libra styled prompt formats with [INST] context [/INST] formatting

Which can easily be adapted from the format specified in the Libra32B repo by replacing alpaca formatting with mixtruct formatting

As always tested in Q8 (not included)

Downloads last month
6
Safetensors
Model size
46.7B params
Tensor type
FP16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Envoid/CybersurferNyandroidLexicat-8x7B

Quantizations
2 models