--- tags: - merge - mergekit - lazymergekit - not-for-all-audiences - nsfw - rp - roleplay - role-play license: llama3 language: - en pipeline_tag: text-generation base_model: - Sao10K/L3-8B-Stheno-v3.2 - ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B - Nitral-AI/Hathor_Stable-v0.2-L3-8B - NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS - Hastagaras/Jamet-8B-L3-MK.V-Blackroot - openlynn/Llama-3-Soliloquy-8B-v2 - NousResearch/Meta-Llama-3-8B-Instruct - turboderp/llama3-turbcat-instruct-8b - VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct - TIGER-Lab/MAmmoTH2-8B-Plus - jondurbin/bagel-8b-v1.0 - abacusai/Llama-3-Smaug-8B - failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 - AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0 - lodrick-the-lafted/Limon-8B - vicgalle/Configurable-Llama-3-8B-v0.3 - Undi95/Llama3-Unholy-8B-OAS - Undi95/Unholy-8B-DPO-OAS --- # L3-Scrambled-Eggs-On-Toast-8B **L3-Scrambled-Eggs-On-Toast-8B** is a role-play model merger **using 18 models** that was made **in 11 merging steps.** The goal is to create both a creative and smart model by using gradients. Each model has their own section in the gradient where they have a larger weight to promote intelligence whereas the rest of the models in the section of the gradient have a small weight to promote creativity. The following models were used as inspiration: * [grimjim/kunoichi-lemon-royale-v3-32K-7B](https://huggingface.co/grimjim/kunoichi-lemon-royale-v3-32K-7B) * [invisietch/EtherealRainbow-v0.3-8B](https://huggingface.co/invisietch/EtherealRainbow-v0.3-8B) * [PJMixers/LLaMa-3-CursedStock-v2.0-8B](https://huggingface.co/PJMixers/LLaMa-3-CursedStock-v2.0-8B) ## Instruct Format Llama 3 ## Settings/Presets ### Instruct/Context Virt-io's [SillyTavern Presets](https://huggingface.co/Virt-io/SillyTavern-Presets/tree/main/Prompts/LLAMA-3/v1.9) is recommended. ### Sampler Settings Here are the current recommended settings for **more creativity** ``` Top K: 60 Min P: 0.035 Rep Pen: 1.05 Rep Pen Range: 2048 Pres Pen: 0.15 Smoothing Factor: 0.25 Dyna Temp: Min Temp: 0.75 Max Temp: 1.5 Expo: 0.85 ``` if you want **more adherence**, then the **Naive preset** is recommended ## Quants Weighted Quants by: - [Lewdiculous](https://huggingface.co/LWDCLS/L3-Scrambled-Eggs-On-Toast-8B-GGUF-IQ-Imatrix-Request) - [mradermacher](https://huggingface.co/mradermacher/L3-Scrambled-Eggs-On-Toast-8B-i1-GGUF) Static Quants by: - [mradermacher](https://huggingface.co/mradermacher/L3-Scrambled-Eggs-On-Toast-8B-GGUF) # Secret Sauce ## Models Used L3-Scrambled-Eggs-On-Toast-8B is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2) * [ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B) * [Nitral-AI/Hathor_Stable-v0.2-L3-8B](https://huggingface.co/Nitral-AI/Hathor_Stable-v0.2-L3-8B) * [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS) * [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot) * [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2) * [NousResearch/Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct) * [turboderp/llama3-turbcat-instruct-8b](https://huggingface.co/turboderp/llama3-turbcat-instruct-8b) * [VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct](https://huggingface.co/VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct) * [TIGER-Lab/MAmmoTH2-8B-Plus](https://huggingface.co/TIGER-Lab/MAmmoTH2-8B-Plus) * [jondurbin/bagel-8b-v1.0](https://huggingface.co/jondurbin/bagel-8b-v1.0) * [abacusai/Llama-3-Smaug-8B](https://huggingface.co/abacusai/Llama-3-Smaug-8B) * [failspy/Meta-Llama-3-8B-Instruct-abliterated-v3](https://huggingface.co/failspy/Meta-Llama-3-8B-Instruct-abliterated-v3) * [AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0](https://huggingface.co/AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0) * [lodrick-the-lafted/Limon-8B](https://huggingface.co/lodrick-the-lafted/Limon-8B) * [vicgalle/Configurable-Llama-3-8B-v0.3](https://huggingface.co/vicgalle/Configurable-Llama-3-8B-v0.3) * [Undi95/Llama3-Unholy-8B-OAS](https://huggingface.co/Undi95/Llama3-Unholy-8B-OAS) * [Undi95/Unholy-8B-DPO-OAS](https://huggingface.co/Undi95/Unholy-8B-DPO-OAS) ## YAML Configs Used The following YAML configs were used to make this mode ### Eggs-and-Bread-RP-pt.1 ```yaml models: - model: Sao10K/L3-8B-Stheno-v3.2 - model: ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B parameters: density: 0.5 weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] - model: Nitral-AI/Hathor_Stable-v0.2-L3-8B parameters: density: 0.5 weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS parameters: density: 0.5 weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot parameters: density: 0.5 weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] - model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 0.5 weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] merge_method: dare_ties base_model: Sao10K/L3-8B-Stheno-v3.2 parameters: normalize: false int8_mask: true dtype: bfloat16 ``` ### Eggs-and-Bread-RP-pt.2 ```yaml models: - model: Sao10K/L3-8B-Stheno-v3.2 - model: ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] - model: Nitral-AI/Hathor_Stable-v0.2-L3-8B parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] - model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: gamma: 0.01 density: 0.9 weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] merge_method: breadcrumbs_ties base_model: Sao10K/L3-8B-Stheno-v3.2 parameters: normalize: false int8_mask: true dtype: bfloat16 ``` ### Egg-and-Bread-RP ```yaml models: - model: Casual-Autopsy/Eggs-and-Bread-RP-pt.1 - model: Casual-Autopsy/Eggs-and-Bread-RP-pt.2 merge_method: slerp base_model: Casual-Autopsy/Eggs-and-Bread-RP-pt.1 parameters: t: - filter: self_attn value: [0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5] - filter: mlp value: [0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5] - value: 0.5 dtype: bfloat16 ``` ### Eggs-and-Bread-IQ-pt.1 ```yaml models: - model: NousResearch/Meta-Llama-3-8B-Instruct - model: turboderp/llama3-turbcat-instruct-8b parameters: density: 0.5 weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] - model: VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct parameters: density: 0.5 weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] - model: TIGER-Lab/MAmmoTH2-8B-Plus parameters: density: 0.5 weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] - model: jondurbin/bagel-8b-v1.0 parameters: density: 0.5 weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] - model: abacusai/Llama-3-Smaug-8B parameters: density: 0.5 weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] merge_method: dare_ties base_model: NousResearch/Meta-Llama-3-8B-Instruct parameters: normalize: false int8_mask: true dtype: bfloat16 ``` ### Eggs-and-Bread-IQ-pt.2 ```yaml models: - model: NousResearch/Meta-Llama-3-8B-Instruct - model: turboderp/llama3-turbcat-instruct-8b parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] - model: VAGOsolutions/Llama-3-SauerkrautLM-8b-Instruct parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] - model: TIGER-Lab/MAmmoTH2-8B-Plus parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] - model: jondurbin/bagel-8b-v1.0 parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] - model: abacusai/Llama-3-Smaug-8B parameters: gamma: 0.01 density: 0.9 weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] merge_method: breadcrumbs_ties base_model: NousResearch/Meta-Llama-3-8B-Instruct parameters: normalize: false int8_mask: true dtype: bfloat16 ``` ### Eggs-and-Bread-IQ ```yaml models: - model: Casual-Autopsy/Eggs-and-Bread-IQ-pt.1 - model: Casual-Autopsy/Eggs-and-Bread-IQ-pt.2 merge_method: slerp base_model: Casual-Autopsy/Eggs-and-Bread-IQ-pt.1 parameters: t: - filter: self_attn value: [0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5] - filter: mlp value: [0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5] - value: 0.5 dtype: bfloat16 ``` ### Eggs-and-Bread-Uncen-pt.1 ```yaml models: - model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 - model: AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0 parameters: density: 0.5 weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] - model: lodrick-the-lafted/Limon-8B parameters: density: 0.5 weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] - model: vicgalle/Configurable-Llama-3-8B-v0.3 parameters: density: 0.5 weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] - model: Undi95/Llama3-Unholy-8B-OAS parameters: density: 0.5 weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] - model: Undi95/Unholy-8B-DPO-OAS parameters: density: 0.5 weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] merge_method: dare_ties base_model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 parameters: normalize: false int8_mask: true dtype: bfloat16 ``` ### Eggs-and-Bread-Uncen-pt.2 ```yaml models: - model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 - model: AwanLLM/Awanllm-Llama-3-8B-Cumulus-v1.0 parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.0825, 0.0825, 0.33] - model: lodrick-the-lafted/Limon-8B parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.0825, 0.33, 0.0825] - model: vicgalle/Configurable-Llama-3-8B-v0.3 parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.0825, 0.33, 0.0825, 0.0825] - model: Undi95/Llama3-Unholy-8B-OAS parameters: gamma: 0.01 density: 0.9 weight: [0.0825, 0.33, 0.0825, 0.0825, 0.0825] - model: Undi95/Unholy-8B-DPO-OAS parameters: gamma: 0.01 density: 0.9 weight: [0.33, 0.0825, 0.0825, 0.0825, 0.0825] merge_method: breadcrumbs_ties base_model: failspy/Meta-Llama-3-8B-Instruct-abliterated-v3 parameters: normalize: false int8_mask: true dtype: bfloat16 ``` ### Eggs-and-Bread-Uncen ```yaml models: - model: Casual-Autopsy/Eggs-and-Bread-Uncen-pt.1 - model: Casual-Autopsy/Eggs-and-Bread-Uncen-pt.2 merge_method: slerp base_model: Casual-Autopsy/Eggs-and-Bread-Uncen-pt.1 parameters: t: - filter: self_attn value: [0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5] - filter: mlp value: [0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5] - value: 0.5 dtype: bfloat16 ``` ### Scrambled-Eggs-On-Toast-1 ```yaml models: - model: Casual-Autopsy/Eggs-and-Bread-RP - model: Casual-Autopsy/Eggs-and-Bread-Uncen merge_method: slerp base_model: Casual-Autopsy/Eggs-and-Bread-RP parameters: t: - value: [0.1, 0.15, 0.2, 0.4, 0.6, 0.4, 0.2, 0.15, 0.1] dtype: bfloat16 ``` ### L3-Scrambled-Eggs-On-Toast-8B ```yaml models: - model: Casual-Autopsy/Scrambled-Eggs-On-Toast-1 - model: Casual-Autopsy/Eggs-and-Bread-IQ merge_method: slerp base_model: Casual-Autopsy/Scrambled-Eggs-On-Toast-1 parameters: t: - value: [0.7, 0.5, 0.3, 0.25, 0.2, 0.25, 0.3, 0.5, 0.7] dtype: bfloat16 ```