tannedbum commited on
Commit
78a6cab
1 Parent(s): 9f5bb00

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +154 -0
README.md ADDED
@@ -0,0 +1,154 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Sao10K/L3-8B-Stheno-v3.2
4
+ - chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO
5
+ - TheDrummer/Llama-3SOME-8B-v2
6
+ - NeverSleep/Llama-3-Lumimaid-8B-v0.1
7
+ - Hastagaras/Jamet-8B-L3-MK.V-Blackroot
8
+ tags:
9
+ - mergekit
10
+ - merge
11
+ - roleplay
12
+ - sillytavern
13
+ - llama3
14
+ - not-for-all-audiences
15
+ license: cc-by-nc-4.0
16
+ language:
17
+ - en
18
+ ---
19
+ ![Nymeria](https://huggingface.co/tannedbum/L3-Nymeria-v2-8B/resolve/main/Nymeria_v2.png??)
20
+
21
+
22
+ - Upgraded SimPO.
23
+ - A touch of 3SOME, Lumimaid and Jamet Blackroot resulting a slightly different prose and wider RP vocab.
24
+ - Leans slightly more on nsfw than the original.
25
+
26
+
27
+ All quants made using imatrix option with dataset provided by bartowski [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
28
+
29
+
30
+ ## SillyTavern
31
+
32
+ ## Text Completion presets
33
+ ```
34
+ temp 0.9
35
+ top_k 30
36
+ top_p 0.75
37
+ min_p 0.2
38
+ rep_pen 1.1
39
+ smooth_factor 0.25
40
+ smooth_curve 1
41
+ ```
42
+ ## Advanced Formatting
43
+
44
+ [Context & Instruct preset by Virt-io](https://huggingface.co/Virt-io/SillyTavern-Presets/tree/main/Prompts/LLAMA-3/v2.0)
45
+
46
+ Instruct Mode: Enabled
47
+
48
+
49
+
50
+ # merge
51
+
52
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
53
+
54
+ This model was merged using the slerp merge method.
55
+
56
+ ### Models Merged
57
+
58
+ The following models were included in the merge:
59
+ * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
60
+ * [chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO](https://huggingface.co/chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO)
61
+ * [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2)
62
+ * [NeverSleep/Llama-3-Lumimaid-8B-v0.1](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1)
63
+ * [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot)
64
+
65
+ ### Configuration
66
+
67
+ The following YAML configuration was used to produce this model:
68
+
69
+ ```yaml
70
+
71
+ slices:
72
+ - sources:
73
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1
74
+ layer_range: [0, 32]
75
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
76
+ layer_range: [0, 32]
77
+ merge_method: slerp
78
+ base_model: NeverSleep/Llama-3-Lumimaid-8B-v0.1
79
+ parameters:
80
+ t:
81
+ - filter: self_attn
82
+ value: [0.7, 0.3, 0.3, 0.3]
83
+ - filter: mlp
84
+ value: [0.3, 0.7, 0.7, 0.7]
85
+ - value: 0.4
86
+ dtype: bfloat16
87
+
88
+ L3-Lumimaid-Jamet-Blackroot-8B
89
+
90
+
91
+ slices:
92
+ - sources:
93
+ - model: tannedbum/L3-Lumimaid-Jamet-Blackroot-8B
94
+ layer_range: [0, 32]
95
+ - model: chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO
96
+ layer_range: [0, 32]
97
+ merge_method: slerp
98
+ base_model: tannedbum/L3-Lumimaid-Jamet-Blackroot-8B
99
+ parameters:
100
+ t:
101
+ - filter: self_attn
102
+ value: [0.3, 0.7, 0.7, 0.7]
103
+ - filter: mlp
104
+ value: [0.7, 0.3, 0.3, 0.3]
105
+ - value: 0.6
106
+ dtype: bfloat16
107
+
108
+ L3-SimPO-Lumimaid-Jamet-Blackroot-8B
109
+
110
+
111
+ slices:
112
+ - sources:
113
+ - model: Sao10K/L3-8B-Stheno-v3.2
114
+ layer_range: [0, 32]
115
+ - model: TheDrummer/Llama-3SOME-8B-v2
116
+ layer_range: [0, 32]
117
+ merge_method: slerp
118
+ base_model: Sao10K/L3-8B-Stheno-v3.2
119
+ parameters:
120
+ t:
121
+ - filter: self_attn
122
+ value: [0.3, 0.3, 0.7, 0.3]
123
+ - filter: mlp
124
+ value: [0.7, 0.7, 0.3, 0.7]
125
+ - value: 0.4
126
+ dtype: bfloat16
127
+
128
+ L3-Stheno-3SOME-8B
129
+
130
+
131
+ slices:
132
+ - sources:
133
+ - model: tannedbum/L3-Stheno-3SOME-8B
134
+ layer_range: [0, 32]
135
+ - model: tannedbum/L3-SimPO-Lumimaid-Jamet-Blackroot-8B
136
+ layer_range: [0, 32]
137
+ merge_method: slerp
138
+ base_model: tannedbum/L3-Stheno-3SOME-8B
139
+ parameters:
140
+ t:
141
+ - filter: self_attn
142
+ value: [0.4, 0.3, 0.3, 0.6]
143
+ - filter: mlp
144
+ value: [0.6, 0.7, 0.7, 0.4]
145
+ - value: 0.4
146
+ dtype: bfloat16
147
+
148
+ L3-Nymeria-v2-8B
149
+
150
+
151
+
152
+ ```
153
+
154
+ Want to support my work ? My Ko-fi page: https://ko-fi.com/tannedbum