Transformers
GGUF
mergekit
Merge
Not-For-All-Audiences
Inference Endpoints
ND911's picture
Create README.md
1f6cd52 verified
---
base_model: []
library_name: transformers
tags:
- mergekit
- merge
- not-for-all-audiences
---
![](eemaid.png)
# Eclectic-Maid-10B
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
A reborn Maid model with many models and several merges. Seems to work well with Alpac Roleplay, ChatML, noromaid settings in Silly Tavern
### Merge Method
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using Eclectic-Maid-10B-p1 as a base.
### Models Merged
Many
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: Eclectic-Maid-10B-p1
# no parameters necessary for base model
- model: Eclectic-Maid-10B-p2
parameters:
weight: 0.50
density: 0.5
- model: Eclectic-Maid-10B-p2
parameters:
weight: 0.50
density: 0.5
merge_method: dare_ties
base_model: Eclectic-Maid-10B-p1
parameters:
int8_mask: true
dtype: bfloat16
name: Eclectic Maid
```
```yaml
name: Maid-Reborn
models:
- model: mistralai/Mistral-7B-Instruct-v0.2
# no parameters necessary for base model
- model: xDAN-AI/xDAN-L1-Chat-RL-v1
parameters:
weight: 0.4
density: 0.8
- model: Undi95/BigL-7B
parameters:
weight: 0.3
density: 0.8
- model: SanjiWatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE
parameters:
weight: 0.2
density: 0.4
- model: NeverSleep/Noromaid-7B-0.4-DPO
parameters:
weight: 0.2
density: 0.4
- model: NSFW_DPO_Noromaid-7B-v2
parameters:
weight: 0.2
density: 0.4
merge_method: dare_ties
base_model: mistralai/Mistral-7B-Instruct-v0.2
parameters:
int8_mask: true
dtype: bfloat16
```
```yaml
models:
- model: Maid-Reborn-7B
# no parameters necessary for base model
- model: localfultonextractor/Erosumika-7B-v2
parameters:
weight: 0.4
density: 0.5
- model: Undi95/BigL-7B
parameters:
weight: 0.3
density: 0.5
- model: Maid-Reborn-7B
parameters:
weight: 0.2
density: 0.4
- model: openchat_openchat-3.5-1210
parameters:
weight: 0.2
density: 0.3
- model: Nexusflow/Starling-LM-7B-beta
parameters:
weight: 0.2
density: 0.3
merge_method: dare_ties
base_model: Maid-Reborn-7B
parameters:
int8_mask: true
dtype: bfloat16
```
```yaml
models:
- model: Maid-Reborn-10B-p1
# no parameters necessary for base model
- model: Maid-Reborn-10B-p2
parameters:
weight: 0.30
density: 0.5
- model: Maid-Reborn-10B-p2
parameters:
weight: 0.30
density: 0.5
merge_method: dare_ties
base_model: Maid-Reborn-10B-p1
parameters:
int8_mask: true
dtype: bfloat16
name: Eclectic Maid
```
```yaml
models:
- model: Maid-Reborn-10B-p2
# no parameters necessary for base model
- model: Maid-Reborn-10B-p1
parameters:
weight: 0.30
density: 0.5
- model: Maid-Reborn-10B-p1
parameters:
weight: 0.30
density: 0.5
merge_method: dare_ties
base_model: Maid-Reborn-10B-p2
parameters:
int8_mask: true
dtype: bfloat16
name: Eclectic Maid
```