AuriAetherwiing commited on
Commit
419f078
1 Parent(s): 55ddaef

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -3
README.md CHANGED
@@ -6,11 +6,51 @@ library_name: transformers
6
  tags:
7
  - mergekit
8
  - merge
 
 
 
 
 
 
 
 
 
 
 
 
 
9
 
10
  ---
11
- # merge-aletheia-7
12
 
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
  ## Merge Details
16
  ### Merge Method
@@ -40,4 +80,4 @@ slices:
40
  model: allura-org/TQ2.5-14B-Neon-v1
41
  - layer_range: [0, 48]
42
  model: allura-org/TQ2.5-14B-Sugarquill-v1
43
- ```
 
6
  tags:
7
  - mergekit
8
  - merge
9
+ license: apache-2.0
10
+ language:
11
+ - en
12
+ ---
13
+ # Qwen2.5-14B Aletheia v1
14
+
15
+ RP/Story hybrid model, merge of Sugarquill and Neon. As with Gemma version, I wanted to preserve Sugarquill's creative spark, while making the model more steerable for RP. It proved to be more difficult this time, but I quite like the result regardless, even if the model is still somewhat temperamental.
16
+
17
+ Should work for both RP and storywriting, either on raw completion or with back-and-forth cowriting in chat mode. Seems to be quite sensitive to low depth instructions and samplers.
18
+
19
+ Thanks to Toasty and Fizz for testing and giving feedback
20
+
21
+ Model was created by Auri.
22
 
23
  ---
 
24
 
25
+ **Notes about merging**
26
+
27
+ It took me 20 something attempts to make this model. TIES didn't work at all, producing broken or nearly broken results every time. SLERP worked much better and after just 3 attempts I got something I like.
28
+ Sugarquill was really prone to overtaking the merge, so I had to reduce it's part a lot, and still model has a lot of influence from it.
29
+
30
+ **Format**
31
+
32
+ Model responds to ChatML instruct formatting, exactly like it's base model.
33
+
34
+ ```
35
+ <|im_start|>system
36
+ {system message}<|im_end|>
37
+ <|im_start|>user
38
+ {user message}<|im_end|>
39
+ <|im_start|>assistant
40
+ {response}<|im_end|>
41
+ ```
42
+
43
+ **Recommended Samplers**
44
+
45
+ This one is a bit of a special snowflake, with special tastes. Those seem to work pretty well:
46
+
47
+ ```
48
+ Temperature - 0.8
49
+ Top-A - 0.3
50
+ TFS - 0.75
51
+ DRY - Multipler 0.8 - Base 1.75 - Allowed length 3 - Range 1024
52
+ ```
53
+
54
 
55
  ## Merge Details
56
  ### Merge Method
 
80
  model: allura-org/TQ2.5-14B-Neon-v1
81
  - layer_range: [0, 48]
82
  model: allura-org/TQ2.5-14B-Sugarquill-v1
83
+ ```