CorticalStack's picture
Update README.md
6e4720f verified
|
raw
history blame
1.32 kB
metadata
license: apache-2.0
tags:
  - merge
  - mergekit
  - mlabonne/NeuralMonarch-7B
  - mlabonne/AlphaMonarch-7B
  - bardsai/jaskier-7b-dpo-v5.6
  - macadeliccc/MBX-7B-v3-DPO

pastiche-crown-clown-7B-dare

pastiche-crown-clown-7B-dare is a DARE merge of the following models using mergekit:

See the paper Language Models are Super Mario: Absorbing Abilities from Homologous Models as a Free Lunch for more on the method.

🧩 Configuration

models:
  - model: mlabonne/NeuralMonarch-7B
    # No parameters necessary for base model
  - model: mlabonne/AlphaMonarch-7B
    parameters:
      density: 0.53
      weight: 0.4
  - model: bardsai/jaskier-7b-dpo-v5.6
    parameters:
      density: 0.53
      weight: 0.3
  - model: macadeliccc/MBX-7B-v3-DPO
    parameters:
      density: 0.53
      weight: 0.3
merge_method: dare_ties
base_model: mlabonne/NeuralMonarch-7B
parameters:
  int8_mask: true
dtype: bfloat16