GGUF
English
Edit model card

GGUF QUANTS OF: https://huggingface.co/Sao10K/Sensualize-Mixtral-bf16


Trained using a randomised subset of Full120k - 60K Samples [Roughly 50M Tokens] + More of my own NSFW Instruct & De-Alignment Data [Roughly 30M Tokens Total]
Total Tokens used for Training: 80M over 1 epoch, over 2xA100s at batch size 5, grad 5 for 12 hours.


Experimental model, trained on mistralai/Mixtral-8x7B-v0.1 using Charles Goddard's ZLoss and Megablocks-based fork of transformers.


Trained with Alpaca format.

### Instruction:
<Prompt>

### Input:
<Insert Context Here>

### Response:

Useful prompt guide: https://rentry.org/mixtralforretards

useful stopping strings:

["\nInput:", "\n[", "\n(", "\n### Input:"] 

Roleplay based model, specifically the ERP type one.

I mean, its kinda alright? I had various testing versions of Mistral 7B and L2 70B, L2 13B, and even Solar with the same dataset and various learning rates, they did much better. MoE tuning kinda meh still.

about gptisms. It's weird. with certain prompts its never there, with some its there. despite the prose of full120k, I never encountered gptslop with mistral, solar or l2 based trains which was why I was confident about this being good. but...

... Enjoy?

Downloads last month
143
GGUF
Unable to determine this model's library. Check the docs .

Dataset used to train Sao10K/Sensualize-Mixtral-GGUF