GGUF
Not-For-All-Audiences
Inference Endpoints
Edit model card

Original model

https://huggingface.co/TheDrummer/Moistral-11B-v2

tldr;

I tried merging Moistral with its base model and got VERY good results with logic & writing. Try them out!

If you do try it out, please let me know your thoughts! Just tell me which percentages you tried and how they performed. The future of Moistral is in your hands!

Moistral 11B v2, dried up

image/gifunmerged model: https://huggingface.co/TheDrummer/Moistral-11B-v2

What is this?

GGUF merges of Moistral 11B v2 and Fimbulvetr v2.

Why merge?

Simply put, the original Moistral v2 can get too moist. It works for some stories, but sometimes you want it to cook.

Which one do I pick?

I'm releasing 5 versions of the merge. The % in the filename represents how much "Fimbulvetr v2" I merged into my finetuned model. The higher the percent, the less moist there is.

Assessment: TBD

Nutshell: All of them are very coherent. Lower percentages = More moist logic & writing.

Added observation: It seems like no matter how moist the situation is, the character reactions are often grounded / realistic. This applies even to 2.5%.

"I found the perfect merge ratio!"

Great! Let me know which one and why. This WILL affect future development.

Downloads last month
82
GGUF
Model size
10.7B params
Architecture
llama
Inference API
Unable to determine this model's library. Check the docs .