|
--- |
|
tags: |
|
- merge |
|
- mergekit |
|
- lazymergekit |
|
- not-for-all-audiences |
|
- nsfw |
|
- rp |
|
- roleplay |
|
- role-play |
|
license: llama3 |
|
language: |
|
- en |
|
pipeline_tag: text-generation |
|
base_model: |
|
- Sao10K/L3-8B-Stheno-v3.2 |
|
- ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B |
|
- Nitral-AI/Hathor_Stable-v0.2-L3-8B |
|
- NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS |
|
- Hastagaras/Jamet-8B-L3-MK.V-Blackroot |
|
- openlynn/Llama-3-Soliloquy-8B-v2 |
|
- NousResearch/Meta-Llama-3-8B-Instruct |
|
- turboderp/llama3-turbcat-instruct-8b |
|
- VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct |
|
- TIGER-Lab/MAmmoTH2-8B-Plus |
|
- jondurbin/bagel-8b-v1.0 |
|
- abacusai/Llama-3-Smaug-8B |
|
- failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 |
|
- AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0 |
|
- lodrick-the-lafted/Limon-8B |
|
- vicgalle/Configurable-Llama-3-8B-v0.3 |
|
- Undi95/Llama3-Unholy-8B-OAS |
|
- Undi95/Unholy-8B-DPO-OAS |
|
--- |
|
|
|
# L3-Scrambled-Eggs-On-Toast-8B |
|
|
|
**L3-Scrambled-Eggs-On-Toast-8B** is a role-play model merger **using 18 models** that was made **in 11 merging steps.** |
|
|
|
The goal is to create both a creative and smart model by using gradients. |
|
Each model has their own section in the gradient where they have a larger weight to promote intelligence whereas the rest of the models in the section of the gradient have a small weight to promote creativity. |
|
|
|
The following models were used as inspiration: |
|
* [grimjim/kunoichi-lemon-royale-v3-32K-7B](https://huggingface.co/grimjim/kunoichi-lemon-royale-v3-32K-7B) |
|
* [invisietch/EtherealRainbow-v0.3-8B](https://huggingface.co/invisietch/EtherealRainbow-v0.3-8B) |
|
* [PJMixers/LLaMa-3-CursedStock-v2.0-8B](https://huggingface.co/PJMixers/LLaMa-3-CursedStock-v2.0-8B) |
|
|
|
## Instruct Format |
|
|
|
Llama 3 |
|
|
|
## Settings/Presets |
|
|
|
### Instruct/Context |
|
|
|
Virt-io's [SillyTavern Presets](https://huggingface.co/Virt-io/SillyTavern-Presets/tree/main/Prompts/LLAMA-3/v1.9) is recommended. |
|
|
|
### Sampler Settings |
|
|
|
Here are the current recommended settings for **more creativity** |
|
``` |
|
Top K: 60 |
|
Min P: 0.035 |
|
Rep Pen: 1.05 |
|
Rep Pen Range: 2048 |
|
Pres Pen: 0.15 |
|
Smoothing Factor: 0.25 |
|
Dyna Temp: |
|
Min Temp: 0.75 |
|
Max Temp: 1.5 |
|
Expo: 0.85 |
|
``` |
|
|
|
if you want **more adherence**, then the **Naive preset** is recommended |
|
|
|
## Quants |
|
|
|
Weighted Quants by: |
|
- [Lewdiculous](https://huggingface.co/LWDCLS/L3-Scrambled-Eggs-On-Toast-8B-GGUF-IQ-Imatrix-Request) |
|
- [mradermacher](https://huggingface.co/mradermacher/L3-Scrambled-Eggs-On-Toast-8B-i1-GGUF) |
|
|
|
Static Quants by: |
|
- [mradermacher](https://huggingface.co/mradermacher/L3-Scrambled-Eggs-On-Toast-8B-GGUF) |
|
|
|
# Secret Sauce |
|
|
|
## Models Used |
|
|
|
L3-Scrambled-Eggs-On-Toast-8B is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): |
|
* [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2) |
|
* [ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B) |
|
* [Nitral-AI/Hathor_Stable-v0.2-L3-8B](https://huggingface.co/Nitral-AI/Hathor_Stable-v0.2-L3-8B) |
|
* [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS) |
|
* [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot) |
|
* [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2) |
|
* [NousResearch/Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct) |
|
* [turboderp/llama3-turbcat-instruct-8b](https://huggingface.co/turboderp/llama3-turbcat-instruct-8b) |
|
* [VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct](https://huggingface.co/VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct) |
|
* [TIGER-Lab/MAmmoTH2-8B-Plus](https://huggingface.co/TIGER-Lab/MAmmoTH2-8B-Plus) |
|
* [jondurbin/bagel-8b-v1.0](https://huggingface.co/jondurbin/bagel-8b-v1.0) |
|
* [abacusai/Llama-3-Smaug-8B](https://huggingface.co/abacusai/Llama-3-Smaug-8B) |
|
* [failspy/Meta-Llama-3-8B-Instruct-abliterated-v3](https://huggingface.co/failspy/Meta-Llama-3-8B-Instruct-abliterated-v3) |
|
* [AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0](https://huggingface.co/AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0) |
|
* [lodrick-the-lafted/Limon-8B](https://huggingface.co/lodrick-the-lafted/Limon-8B) |
|
* [vicgalle/Configurable-Llama-3-8B-v0.3](https://huggingface.co/vicgalle/Configurable-Llama-3-8B-v0.3) |
|
* [Undi95/Llama3-Unholy-8B-OAS](https://huggingface.co/Undi95/Llama3-Unholy-8B-OAS) |
|
* [Undi95/Unholy-8B-DPO-OAS](https://huggingface.co/Undi95/Unholy-8B-DPO-OAS) |
|
|
|
## YAML Configs Used |
|
|
|
The following YAML configs were used to make this mode |
|
|
|
### Eggs-and-Bread-RP-pt.1 |
|
|
|
```yaml |
|
models: |
|
- model: Sao10K/L3-8B-Stheno-v3.2 |
|
- model: ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B |
|
parameters: |
|
density: 0.5 |
|
weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] |
|
- model: Nitral-AI/Hathor_Stable-v0.2-L3-8B |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] |
|
- model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] |
|
- model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] |
|
- model: openlynn/Llama-3-Soliloquy-8B-v2 |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] |
|
merge_method: dare_ties |
|
base_model: Sao10K/L3-8B-Stheno-v3.2 |
|
parameters: |
|
normalize: false |
|
int8_mask: true |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Eggs-and-Bread-RP-pt.2 |
|
|
|
```yaml |
|
models: |
|
- model: Sao10K/L3-8B-Stheno-v3.2 |
|
- model: ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] |
|
- model: Nitral-AI/Hathor_Stable-v0.2-L3-8B |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] |
|
- model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] |
|
- model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] |
|
- model: openlynn/Llama-3-Soliloquy-8B-v2 |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] |
|
merge_method: breadcrumbs_ties |
|
base_model: Sao10K/L3-8B-Stheno-v3.2 |
|
parameters: |
|
normalize: false |
|
int8_mask: true |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Egg-and-Bread-RP |
|
|
|
```yaml |
|
models: |
|
- model: Casual-Autopsy/Eggs-and-Bread-RP-pt.1 |
|
- model: Casual-Autopsy/Eggs-and-Bread-RP-pt.2 |
|
merge_method: slerp |
|
base_model: Casual-Autopsy/Eggs-and-Bread-RP-pt.1 |
|
parameters: |
|
t: |
|
- filter: self_attn |
|
value: [0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5] |
|
- filter: mlp |
|
value: [0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5] |
|
- value: 0.5 |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Eggs-and-Bread-IQ-pt.1 |
|
|
|
```yaml |
|
models: |
|
- model: NousResearch/Meta-Llama-3-8B-Instruct |
|
- model: turboderp/llama3-turbcat-instruct-8b |
|
parameters: |
|
density: 0.5 |
|
weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] |
|
- model: VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] |
|
- model: TIGER-Lab/MAmmoTH2-8B-Plus |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] |
|
- model: jondurbin/bagel-8b-v1.0 |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] |
|
- model: abacusai/Llama-3-Smaug-8B |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] |
|
merge_method: dare_ties |
|
base_model: NousResearch/Meta-Llama-3-8B-Instruct |
|
parameters: |
|
normalize: false |
|
int8_mask: true |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Eggs-and-Bread-IQ-pt.2 |
|
|
|
```yaml |
|
models: |
|
- model: NousResearch/Meta-Llama-3-8B-Instruct |
|
- model: turboderp/llama3-turbcat-instruct-8b |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] |
|
- model: VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] |
|
- model: TIGER-Lab/MAmmoTH2-8B-Plus |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] |
|
- model: jondurbin/bagel-8b-v1.0 |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] |
|
- model: abacusai/Llama-3-Smaug-8B |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] |
|
merge_method: breadcrumbs_ties |
|
base_model: NousResearch/Meta-Llama-3-8B-Instruct |
|
parameters: |
|
normalize: false |
|
int8_mask: true |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Eggs-and-Bread-IQ |
|
|
|
```yaml |
|
models: |
|
- model: Casual-Autopsy/Eggs-and-Bread-IQ-pt.1 |
|
- model: Casual-Autopsy/Eggs-and-Bread-IQ-pt.2 |
|
merge_method: slerp |
|
base_model: Casual-Autopsy/Eggs-and-Bread-IQ-pt.1 |
|
parameters: |
|
t: |
|
- filter: self_attn |
|
value: [0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5] |
|
- filter: mlp |
|
value: [0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5] |
|
- value: 0.5 |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Eggs-and-Bread-Uncen-pt.1 |
|
|
|
```yaml |
|
models: |
|
- model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 |
|
- model: AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0 |
|
parameters: |
|
density: 0.5 |
|
weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] |
|
- model: lodrick-the-lafted/Limon-8B |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] |
|
- model: vicgalle/Configurable-Llama-3-8B-v0.3 |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] |
|
- model: Undi95/Llama3-Unholy-8B-OAS |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] |
|
- model: Undi95/Unholy-8B-DPO-OAS |
|
parameters: |
|
density: 0.5 |
|
weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] |
|
merge_method: dare_ties |
|
base_model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 |
|
parameters: |
|
normalize: false |
|
int8_mask: true |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Eggs-and-Bread-Uncen-pt.2 |
|
|
|
```yaml |
|
models: |
|
- model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 |
|
- model: AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0 |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] |
|
- model: lodrick-the-lafted/Limon-8B |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] |
|
- model: vicgalle/Configurable-Llama-3-8B-v0.3 |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] |
|
- model: Undi95/Llama3-Unholy-8B-OAS |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] |
|
- model: Undi95/Unholy-8B-DPO-OAS |
|
parameters: |
|
gamma: 0.01 |
|
density: 0.9 |
|
weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] |
|
merge_method: breadcrumbs_ties |
|
base_model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 |
|
parameters: |
|
normalize: false |
|
int8_mask: true |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Eggs-and-Bread-Uncen |
|
|
|
```yaml |
|
models: |
|
- model: Casual-Autopsy/Eggs-and-Bread-Uncen-pt.1 |
|
- model: Casual-Autopsy/Eggs-and-Bread-Uncen-pt.2 |
|
merge_method: slerp |
|
base_model: Casual-Autopsy/Eggs-and-Bread-Uncen-pt.1 |
|
parameters: |
|
t: |
|
- filter: self_attn |
|
value: [0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5] |
|
- filter: mlp |
|
value: [0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5] |
|
- value: 0.5 |
|
dtype: bfloat16 |
|
``` |
|
|
|
### Scrambled-Eggs-On-Toast-1 |
|
|
|
```yaml |
|
models: |
|
- model: Casual-Autopsy/Eggs-and-Bread-RP |
|
- model: Casual-Autopsy/Eggs-and-Bread-Uncen |
|
merge_method: slerp |
|
base_model: Casual-Autopsy/Eggs-and-Bread-RP |
|
parameters: |
|
t: |
|
- value: [0.1, 0.15, 0.2, 0.4, 0.6, 0.4, 0.2, 0.15, 0.1] |
|
dtype: bfloat16 |
|
``` |
|
|
|
### L3-Scrambled-Eggs-On-Toast-8B |
|
|
|
```yaml |
|
models: |
|
- model: Casual-Autopsy/Scrambled-Eggs-On-Toast-1 |
|
- model: Casual-Autopsy/Eggs-and-Bread-IQ |
|
merge_method: slerp |
|
base_model: Casual-Autopsy/Scrambled-Eggs-On-Toast-1 |
|
parameters: |
|
t: |
|
- value: [0.7, 0.5, 0.3, 0.25, 0.2, 0.25, 0.3, 0.5, 0.7] |
|
dtype: bfloat16 |
|
``` |