Deris-SSS-7B / README.md
Virt-io's picture
Upload folder using huggingface_hub
2010bee verified
|
raw
history blame
2.83 kB
---
library_name: transformers
tags:
- mergekit
- merge
---
# Tips
SillyTavern presets in presets folder.
Model has formatting issues when using asterisks. Recommended to use novel like formatting (only use quotes)
System prompt can be improved, help welcomed.
Model seems to take characters too seriously, if you find it too stubborn regenerate reply or edit it. It should comply after.
# Deris-SSS
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the SLERP merge method.
### Models Merged
The following models were included in the merge:
* Deris-v2
* SmartyPants-step2
### Configuration
The following YAML configuration was used to produce this model:
### Deris-SSS
Final merge combine the smart models with the unhinged ones
```yaml
slices:
- sources:
- model: ./Mergekit/Deris-v2
layer_range: [0, 32]
- model: ./Mergekit/SmartyPants-step2
layer_range: [0, 32]
merge_method: slerp
base_model: ./Mergekit/Deris-v2
parameters:
t:
- filter: self_attn
value: [0, 0.5, 0.3, 0.7, 1]
- filter: mlp
value: [1, 0.5, 0.7, 0.3, 0]
- value: 0.5
dtype: float16
```
### Deris-v2
Combine a bit of Datura_7B into Eris_Floramix_DPO_7B
Reason Datura is extremrly unhinged :) more so than Eris
```
slices:
- sources:
- model: ChaoticNeutrals/Eris_Floramix_DPO_7B
layer_range: [0, 32]
- model: ResplendentAI/Datura_7B
layer_range: [0, 32]
merge_method: slerp
base_model: ChaoticNeutrals/Eris_Floramix_DPO_7B
parameters:
t:
- filter: self_attn
value: [0, 0.20, 0.15, 0.25, 0.35]
- filter: mlp
value: [0.35, 0.20, 0.25, 0.15, 0]
- value: 0.20
dtype: float16
```
### SmartyPants-step1
Combine OMJ into Einstein
Reason Einstein looks interesting and OMJ was a high ranking model
```
slices:
- sources:
- model: Weyaxi/Einstein-v4-7B
layer_range: [0, 32]
- model: eren23/ogno-monarch-jaskier-merge-7b-OH-PREF-DPO
layer_range: [0, 32]
merge_method: slerp
base_model: Weyaxi/Einstein-v4-7B
parameters:
t:
- filter: self_attn
value: [0, 0.45, 0.3, 0.55, 0.65]
- filter: mlp
value: [0.65, 0.45, 0.55, 0.3, 0]
- value: 0.45
dtype: float16
```
### SmartyPants-step2
Combine Smarty pants into FuseChat-VaRM
Reason IDK I just like FuseChat-VaRM
```
slices:
- sources:
- model: FuseAI/FuseChat-7B-VaRM
layer_range: [0, 32]
- model: ./Mergekit/SmartyPants-step1
layer_range: [0, 32]
merge_method: slerp
base_model: FuseAI/FuseChat-7B-VaRM
parameters:
t:
- filter: self_attn
value: [0, 0.45, 0.3, 0.55, 0.65]
- filter: mlp
value: [0.65, 0.45, 0.55, 0.3, 0]
- value: 0.45
dtype: float16
```