asiansoul commited on
Commit
01465da
1 Parent(s): b241208

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +110 -5
README.md CHANGED
@@ -1,5 +1,110 @@
1
- ---
2
- license: other
3
- license_name: other
4
- license_link: LICENSE
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - beomi/Llama-3-KoEn-8B-Instruct-preview
4
+ - asiansoul/Joah-Llama-3-MAAL-MLP-KoEn-8B-Reborn
5
+ - NousResearch/Hermes-2-Pro-Llama-3-8B
6
+ - saltlux/Ko-Llama3-Luxia-8B
7
+ - defog/llama-3-sqlcoder-8b
8
+ - Locutusque/llama-3-neural-chat-v2.2-8B
9
+ - rombodawg/Llama-3-8B-Instruct-Coder
10
+ - NousResearch/Meta-Llama-3-8B-Instruct
11
+ - aaditya/Llama3-OpenBioLLM-8B
12
+ - rombodawg/Llama-3-8B-Base-Coder-v3.5-10k
13
+ - cognitivecomputations/dolphin-2.9.1-llama-3-8b
14
+ - abacusai/Llama-3-Smaug-8B
15
+ - NousResearch/Meta-Llama-3-8B
16
+ library_name: transformers
17
+ tags:
18
+ - mergekit
19
+ - merge
20
+
21
+ ---
22
+ # Joah-Llama-3-MAAL-MLP-KoEn-8B-Reborn-V1
23
+
24
+ ## Merge Details
25
+ ### Merge Method
26
+
27
+ This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) as a base.
28
+
29
+ ### Models Merged
30
+
31
+ The following models were included in the merge:
32
+ * [beomi/Llama-3-KoEn-8B-Instruct-preview](https://huggingface.co/beomi/Llama-3-KoEn-8B-Instruct-preview)
33
+ * [asiansoul/Joah-Llama-3-MAAL-MLP-KoEn-8B-Reborn](https://huggingface.co/asiansoul/Joah-Llama-3-MAAL-MLP-KoEn-8B-Reborn)
34
+ * [NousResearch/Hermes-2-Pro-Llama-3-8B](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B)
35
+ * [saltlux/Ko-Llama3-Luxia-8B](https://huggingface.co/saltlux/Ko-Llama3-Luxia-8B)
36
+ * [defog/llama-3-sqlcoder-8b](https://huggingface.co/defog/llama-3-sqlcoder-8b)
37
+ * [Locutusque/llama-3-neural-chat-v2.2-8B](https://huggingface.co/Locutusque/llama-3-neural-chat-v2.2-8B)
38
+ * [rombodawg/Llama-3-8B-Instruct-Coder](https://huggingface.co/rombodawg/Llama-3-8B-Instruct-Coder)
39
+ * [NousResearch/Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct)
40
+ * [aaditya/Llama3-OpenBioLLM-8B](https://huggingface.co/aaditya/Llama3-OpenBioLLM-8B)
41
+ * [rombodawg/Llama-3-8B-Base-Coder-v3.5-10k](https://huggingface.co/rombodawg/Llama-3-8B-Base-Coder-v3.5-10k)
42
+ * [cognitivecomputations/dolphin-2.9.1-llama-3-8b](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-8b)
43
+ * [abacusai/Llama-3-Smaug-8B](https://huggingface.co/abacusai/Llama-3-Smaug-8B)
44
+
45
+ ### Configuration
46
+
47
+ The following YAML configuration was used to produce this model:
48
+
49
+ ```yaml
50
+ models:
51
+ - model: NousResearch/Meta-Llama-3-8B
52
+ # Base model providing a general foundation without specific parameters
53
+ - model: NousResearch/Meta-Llama-3-8B-Instruct
54
+ parameters:
55
+ density: 0.65
56
+ weight: 0.25
57
+
58
+ - model: asiansoul/Joah-Llama-3-MAAL-MLP-KoEn-8B-Reborn
59
+ parameters:
60
+ density: 0.6
61
+ weight: 0.2
62
+ - model: beomi/Llama-3-KoEn-8B-Instruct-preview
63
+ parameters:
64
+ density: 0.55
65
+ weight: 0.125
66
+
67
+ - model: saltlux/Ko-Llama3-Luxia-8B
68
+ parameters:
69
+ density: 0.55
70
+ weight: 0.125
71
+ - model: cognitivecomputations/dolphin-2.9.1-llama-3-8b
72
+ parameters:
73
+ density: 0.55
74
+ weight: 0.05
75
+ - model: aaditya/Llama3-OpenBioLLM-8B
76
+ parameters:
77
+ density: 0.55
78
+ weight: 0.05
79
+
80
+ - model: rombodawg/Llama-3-8B-Instruct-Coder
81
+ parameters:
82
+ density: 0.55
83
+ weight: 0.05
84
+ - model: rombodawg/Llama-3-8B-Base-Coder-v3.5-10k
85
+ parameters:
86
+ density: 0.55
87
+ weight: 0.05
88
+ - model: defog/llama-3-sqlcoder-8b
89
+ parameters:
90
+ density: 0.55
91
+ weight: 0.05
92
+ - model: Locutusque/llama-3-neural-chat-v2.2-8B
93
+ parameters:
94
+ density: 0.55
95
+ weight: 0.05
96
+ - model: NousResearch/Hermes-2-Pro-Llama-3-8B
97
+ parameters:
98
+ density: 0.55
99
+ weight: 0.05
100
+
101
+ - model: abacusai/Llama-3-Smaug-8B
102
+ parameters:
103
+ density: 0.55
104
+ weight: 0.05
105
+ merge_method: dare_ties
106
+ base_model: NousResearch/Meta-Llama-3-8B
107
+ parameters:
108
+ int8_mask: true
109
+ dtype: bfloat16
110
+ ```