Hermyale-stack-90B

It's a stack merge meme model made from Hermes 3 and Euryale 2.2. You should use it for roleplay and creative writing AND PROBABLY NOTHING ELSE (but hey, it's your funeral)

STACK MERGE DISCLAIMER

yes it's just a stack merge, no I didn't do any additional pretraining, no stack merges don't make the model smarter, yes they harm its ability to do complex logical tasks, yes they introduce some weird behaviors and unexpected mistakes, no they don't make the model sentient, no you shouldn't post on twitter about how adding a few layers turned it into agi, etc. etc.

That said, it does feel unique and fun to use. If you're the type of person who's drowning in VRAM would rather have some more variety at the expense of needing to make a few manual edits to clean up mistakes, give it a try.

Format

ChatML

Samplers

Because stack merges introduce some unexpected noise to the model, I recommend higher min p than normal. I've been getting good results with min_p 0.1 -> temp 1.1 (I usually prefer something like min_p 0.03-0.05 -> temp 0.7-0.9, adjust according to taste). Add your favorite anti-repetition sampler as needed.

Merge Rationale

The config is below:

slices:
  - sources:
    - model: ../Hermes-3-Llama-3.1-70B
      layer_range: [0, 21]
  - sources:
    - model: ../L3.1-70B-Euryale-v2.2
      layer_range: [16, 36]
  - sources:
    - model: ../Hermes-3-Llama-3.1-70B
      layer_range: [30, 50]
  - sources:
    - model: ../L3.1-70B-Euryale-v2.2
      layer_range: [40, 64]
  - sources:
    - model: ../Hermes-3-Llama-3.1-70B
      layer_range: [60, 80]

tokenizer_source: ../Hermes-3-Llama-3.1-70B 
merge_method: passthrough
dtype: bfloat16

You will notice these layer ranges are less uniform and logically-spaced than most stack merges. That's because I tested a bunch of versions of this with different slices to see which ones felt best before eventually arriving at this config, which seems to strike the best balance between fun creativity, coherence, and size.

Random theorizing:

  • Including a huge number of layers from both models is not necessary to achieve the beneficial aspects of stack merges (e.g. I don't think a 120B variant of this would be any better)
  • I avoid taking tons of smaller slices as some amount of continuity is beneficial, but too much and you aren't getting that stack merge je ne sais quoi.
  • I try to focus most of the layer overlap on the middle layers, as these seem to be the most composable and have the best bang for your buck when stack merging.
  • Messing too much with the later layers seems to be almost entirely pointless.
  • Messing too much with the early layers can be very detrimental to coherence (but also, letting a single model get too far through its early layers uninterrupted can reduce the Fun Factor of the resulting merge)

All credit goes to the original finetuners, I'm just some dummy who can write mergekit configs.

:*

Downloads last month
15
Safetensors
Model size
91.9B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for mergesloppa123123/Hermyale-stack-90B