tannedbum commited on
Commit
c1efedd
·
verified ·
1 Parent(s): 799b246

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - princeton-nlp/gemma-2-9b-it-SimPO
4
+ - TheDrummer/Gemmasutra-9B-v1
5
+ tags:
6
+ - mergekit
7
+ - merge
8
+ - roleplay
9
+ - sillytavern
10
+ - gemma2
11
+ - not-for-all-audiences
12
+ license: cc-by-nc-4.0
13
+ language:
14
+ - en
15
+ ---
16
+
17
+ All quants made using imatrix option with dataset provided by bartowski [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
18
+
19
+
20
+ ## SillyTavern
21
+
22
+ ## Text Completion presets
23
+ ```
24
+ temp 0.9
25
+ top_k 30
26
+ top_p 0.75
27
+ min_p 0.2
28
+ rep_pen 1.1
29
+ smooth_factor 0.25
30
+ smooth_curve 1
31
+ ```
32
+ ## Advanced Formatting
33
+
34
+
35
+ Context & Instruct Presets for Gemma [Here](https://huggingface.co/tannedbum/ST-Presets/tree/main) IMPORTANT !
36
+
37
+ Instruct Mode: Enabled
38
+
39
+
40
+
41
+
42
+
43
+
44
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
45
+
46
+ ### Merge Method
47
+
48
+ This model was merged using the SLERP merge method.
49
+
50
+ ### Models Merged
51
+
52
+ The following models were included in the merge:
53
+ * [princeton-nlp/gemma-2-9b-it-SimPO](https://huggingface.co/princeton-nlp/gemma-2-9b-it-SimPO)
54
+ * [TheDrummer/Gemmasutra-9B-v1](https://huggingface.co/TheDrummer/Gemmasutra-9B-v1)
55
+
56
+ ### Configuration
57
+
58
+ The following YAML configuration was used to produce this model:
59
+
60
+ ```yaml
61
+ slices:
62
+ - sources:
63
+ - model: TheDrummer/Gemmasutra-9B-v1
64
+ layer_range: [0, 42]
65
+ - model: princeton-nlp/gemma-2-9b-it-SimPO
66
+ layer_range: [0, 42]
67
+ merge_method: slerp
68
+ base_model: TheDrummer/Gemmasutra-9B-v1
69
+ parameters:
70
+ t:
71
+ - filter: self_attn
72
+ value: [0.2, 0.4, 0.6, 0.2, 0.4]
73
+ - filter: mlp
74
+ value: [0.8, 0.6, 0.4, 0.8, 0.6]
75
+ - value: 0.4
76
+ dtype: bfloat16
77
+
78
+
79
+
80
+
81
+
82
+ ```