models: | |
- model: mlabonne/NeuralHermes-2.5-Mistral-7B # no parameters necessary for base model | |
- model: VAGOsolutions/SauerkrautLM-7b-HerO | |
parameters: | |
density: 0.3 # fraction of weights in differences from the base model to retain | |
weight: # weight gradient | |
- filter: mlp | |
value: 0.5 | |
- value: 0 | |
- model: cognitivecomputations/dolphin-2.8-mistral-7b-v02 | |
parameters: | |
density: 0.5 | |
weight: 0.4 | |
merge_method: ties | |
base_model: mlabonne/NeuralHermes-2.5-Mistral-7B | |
parameters: | |
normalize: true | |
int8_mask: true | |
dtype: float16 | |