Statuo commited on
Commit
0c0288d
1 Parent(s): 93b211f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -2
README.md CHANGED
@@ -6,8 +6,21 @@ library_name: transformers
6
  tags:
7
  - mergekit
8
  - merge
9
-
10
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  # merge
12
 
13
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
@@ -36,4 +49,4 @@ base_model: KatyTheCutie/LemonadeRP-4.5.3
36
  dtype: bfloat16
37
  parameters:
38
  t: [0, 0.5, 1, 0.5, 0] # V shaped curve: Hermes for input & output, WizardMath in the middle layers
39
- ```
 
6
  tags:
7
  - mergekit
8
  - merge
9
+ license: cc-by-nc-4.0
10
  ---
11
+ # Intent
12
+ The intent was to combine the excellent LemonadeRP-4.5.3 with WizardLM-2 in order to produce more effective uncensored content. While WizardLM-2 wouldn't balk at uncensored content, it would still falter in actually producing it whereas LemonadeRP didn't have this issue. The results are pretty good imo. There's a problem that if your response length is too long it will start to speak for the user but those usually disappear on swipes.
13
+
14
+ I had originally not intended to release this model and instead keep it private. It's my first foray into doing merges at all and I didn't want to release a subpar model. However, after encouragement I've decided to unprivate it. Hope you all get some enjoyment out of it.
15
+
16
+ # Prompt - Alpaca
17
+
18
+ Using the Alpaca prompt seems to get good results.
19
+
20
+ # Context Size - 8192
21
+
22
+ Haven't tested beyond this. Usual rule of thumb is that once you get up to 12k your responses tend to become less coherent and 16k is where things just devolve completely.
23
+
24
  # merge
25
 
26
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
49
  dtype: bfloat16
50
  parameters:
51
  t: [0, 0.5, 1, 0.5, 0] # V shaped curve: Hermes for input & output, WizardMath in the middle layers
52
+ ```