fp16: https://huggingface.co/Sao10K/Typhon-Mixtral-v1
exl2: https://huggingface.co/Sao10K/Typhon-Mixtral-v1-exl2
Typhon - A Custom Experimental Mixtral Merge
An experimental Merge I tried for fun. Honestly did not expect it to work for Mixtral at all considering how its an MoE and the gates and all would be fucked by this custom merge.
From my testing it was able to handle SFW <--> NSFW scenarios fine, handle 1st and 3rd person roleplays fine, and seemed fairly smart.
It did pretty well for non NSFW tasks so that's a win.
Recipe Below:
base_model: mistralai/Mixtral-8x7B-v0.1
models:
- model: mistralai/Mixtral-8x7B-v0.1
# no parameters necessary for base model
- model: smelborp/MixtralOrochi8x7B
parameters:
weight: 0.30
density: 0.47
- model: notstoic/Nous-Hermes-2-Mixtruct-v0.1-8x7B-DPO-DARE_TIES
parameters:
weight: 0.31
density: 0.56
- model: Sao10K/Solstice-Mixtral-v1
parameters:
weight: 0.36
density: 0.64
- model: Sao10K/Frostwind-Mixtral-v1
parameters:
weight: 0.22
density: 0.44
- model: KoboldAI/Mixtral-8x7B-Holodeck-v1
parameters:
weight: 0.21
density: 0.36
merge_method: dare_ties
dtype: bfloat16
- Downloads last month
- 954
Unable to determine this model's library. Check the
docs
.