--- base_model: - cgato/L3-TheSpice-8b-v0.8.3 - NousResearch/Hermes-2-Pro-Llama-3-8B - openlynn/Llama-3-Soliloquy-8B-v2 library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [openlynn/Llama-3-Soliloquy-8B-v2](https://huggingface.co/openlynn/Llama-3-Soliloquy-8B-v2) as a base. ### Models Merged The following models were included in the merge: * [cgato/L3-TheSpice-8b-v0.8.3](https://huggingface.co/cgato/L3-TheSpice-8b-v0.8.3) * [NousResearch/Hermes-2-Pro-Llama-3-8B](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: openlynn/Llama-3-Soliloquy-8B-v2 dtype: bfloat16 merge_method: dare_ties parameters: int8_mask: 1.0 normalize: 0.0 slices: - sources: - layer_range: [0, 4] model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 1.0 weight: 0.6861808716092435 - layer_range: [0, 4] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 0.6628290134113985 weight: 0.5815923052193855 - layer_range: [0, 4] model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 1.0 weight: 0.5113886163963061 - sources: - layer_range: [4, 8] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 0.892655547455918 weight: 0.038732602391021484 - layer_range: [4, 8] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 1.0 weight: 0.1982145486303527 - layer_range: [4, 8] model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 1.0 weight: 0.6843011350690802 - sources: - layer_range: [8, 12] model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 0.7817511027396784 weight: 0.13053333213489704 - layer_range: [8, 12] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 0.6963703515864826 weight: 0.20525481492667985 - layer_range: [8, 12] model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 0.6983086326765777 weight: 0.5843953969574106 - sources: - layer_range: [12, 16] model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 0.9632895768462915 weight: 0.2101146706607748 - layer_range: [12, 16] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 0.597557434542081 weight: 0.6728172621848589 - layer_range: [12, 16] model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 0.756263557607837 weight: 0.2581423726361908 - sources: - layer_range: [16, 20] model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 1.0 weight: 0.2116035543552448 - layer_range: [16, 20] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 1.0 weight: 0.22654226422958418 - layer_range: [16, 20] model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 0.8925914810507647 weight: 0.42243766315440867 - sources: - layer_range: [20, 24] model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 0.7697608089825734 weight: 0.1535118632140203 - layer_range: [20, 24] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 0.9886758076773643 weight: 0.3305040603868546 - layer_range: [20, 24] model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 1.0 weight: 0.40670083428654535 - sources: - layer_range: [24, 28] model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 1.0 weight: 0.4542810478500622 - layer_range: [24, 28] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 0.8330662483310117 weight: 0.2587495367324508 - layer_range: [24, 28] model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 0.9845313983551542 weight: 0.40378452705975915 - sources: - layer_range: [28, 32] model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 1.0 weight: 0.2951962192288415 - layer_range: [28, 32] model: cgato/L3-TheSpice-8b-v0.8.3 parameters: density: 0.960315594933433 weight: 0.13142971773782525 - layer_range: [28, 32] model: openlynn/Llama-3-Soliloquy-8B-v2 parameters: density: 1.0 weight: 0.30838472094518804 ```