Good Evening.
New model mix because I got frustrated of dealing with wrangling with parameters and I chalked it up to a "Seems like a model issue".
Going back to merge stock since i dont feel like experimenting and want to try something that vibes well out of the box.
Merge Details
Merge Method
This model was merged using the Model Stock merge method using Llama-3.3-70B-Instruct as a base.
Models Merged
The following models were included in the merge:
- KaraKaraWitch/Llama-MiraiFanfare-2-3.3-70B
- Undi95/Sushi-v1.4
- Nohobby/L3.3-Prikol-70B-v0.2
- Sao10K/L3.3-70B-Euryale-v2.3
- TheDrummer/Anubis-70B-v1
- EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
- nitky/Llama-3.3-SuperSwallowX-70B-Instruct-v0.1
- Blackroot/Mirai-3.0-70B
- Sao10K/70B-L3.3-Cirrus-x1
Configuration
The following YAML configuration was used to produce this model:
models:
- model: Blackroot/Mirai-3.0-70B
- model: EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1
- model: TheDrummer/Anubis-70B-v1
- model: Sao10K/L3.3-70B-Euryale-v2.3
- model: Sao10K/70B-L3.3-Cirrus-x1
- model: nitky/Llama-3.3-SuperSwallowX-70B-Instruct-v0.1
- model: KaraKaraWitch/Llama-MiraiFanfare-2-3.3-70B
- model: Undi95/Sushi-v1.4
- model: Nohobby/L3.3-Prikol-70B-v0.2
merge_method: model_stock
base_model: Llama-3.3-70B-Instruct
parameters:
normalize: true
dtype: bfloat16