Doctor-Shotgun commited on
Commit
8e8da0d
1 Parent(s): 15e6f3e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +63 -12
README.md CHANGED
@@ -2,14 +2,17 @@
2
  library_name: peft
3
  tags:
4
  - generated_from_trainer
 
 
5
  model-index:
6
  - name: volume/limarp-70b-qlora
7
  results: []
 
 
 
 
8
  ---
9
 
10
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
- should probably proofread and complete it, then remove this comment. -->
12
-
13
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
14
  <details><summary>See axolotl config</summary>
15
 
@@ -92,21 +95,73 @@ special_tokens:
92
 
93
  </details><br>
94
 
95
- # volume/limarp-70b-qlora
 
 
 
 
96
 
97
- This model was trained from scratch on the None dataset.
98
 
99
  ## Model description
100
 
101
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
102
 
103
  ## Intended uses & limitations
104
 
105
- More information needed
106
 
107
  ## Training and evaluation data
108
 
109
- More information needed
110
 
111
  ## Training procedure
112
 
@@ -124,10 +179,6 @@ The following hyperparameters were used during training:
124
  - lr_scheduler_warmup_steps: 10
125
  - num_epochs: 2
126
 
127
- ### Training results
128
-
129
-
130
-
131
  ### Framework versions
132
 
133
  - PEFT 0.7.2.dev0
 
2
  library_name: peft
3
  tags:
4
  - generated_from_trainer
5
+ - llama
6
+ - llama 2
7
  model-index:
8
  - name: volume/limarp-70b-qlora
9
  results: []
10
+ datasets:
11
+ - lemonilia/LimaRP
12
+ language:
13
+ - en
14
  ---
15
 
 
 
 
16
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
17
  <details><summary>See axolotl config</summary>
18
 
 
95
 
96
  </details><br>
97
 
98
+ # limarp-miqu-1-70b-qlora
99
+
100
+ Experimental limarp qlora trained at 16384 ctx length (greater than size of the longest limarp sample when tokenized via mistral's tokenizer) on the fixed dequantized miqu-1-70b model by 152334H.
101
+
102
+ I wasn't particularly happy with the results I got when I tried applying the lora at varying weights to the miqu-1-70b model. It's possible that this is related to the fact that the model was dequantized from Q5_K_M GGUF, or perhaps due to it already being an instruct-tuned model.
103
 
104
+ However, I decided to go ahead and release this in case someone else finds a use for it.
105
 
106
  ## Model description
107
 
108
+ The intended prompt format is the Alpaca instruction format of LimaRP v3:
109
+ ```
110
+ ### Instruction:
111
+ Character's Persona: {bot character description}
112
+
113
+ User's Persona: {user character description}
114
+
115
+ Scenario: {what happens in the story}
116
+
117
+ Play the role of Character. Taking the above information into consideration, you must engage in a roleplaying chat with User below this line. Do not write dialogues and narration for User.
118
+
119
+ ### Input:
120
+ User: {utterance}
121
+
122
+ ### Response:
123
+ Character: {utterance}
124
+
125
+ ### Input:
126
+ User: {utterance}
127
+
128
+ ### Response:
129
+ Character: {utterance}
130
+
131
+ (etc.)
132
+ ```
133
+ Inspired by the previously named "Roleplay" preset in SillyTavern, with this version of LimaRP it is possible to append a length modifier to the response instruction sequence, like this:
134
+
135
+ ```
136
+ ### Input
137
+ User: {utterance}
138
+
139
+ ### Response: (length = medium)
140
+ Character: {utterance}
141
+ ```
142
+
143
+ This has an immediately noticeable effect on bot responses. The lengths using during training are:
144
+ `micro`, `tiny`, `short`, `medium`, `long`, `massive`, `huge`, `enormous`, `humongous`, `unlimited`.
145
+ **The recommended starting length is medium**. Keep in mind that the AI can ramble or impersonate
146
+ the user with very long messages.
147
+
148
+ The length control effect is reproducible, but the messages will not necessarily follow
149
+ lengths very precisely, rather follow certain ranges on average, as seen in this table
150
+ with data from tests made with one reply at the beginning of the conversation:
151
+
152
+ ![lengths](https://i.imgur.com/2WXGgaV.png)
153
+
154
+ Response length control appears to work well also deep into the conversation. **By omitting
155
+ the modifier, the model will choose the most appropriate response length** (although it might
156
+ not necessarily be what the user desires).
157
 
158
  ## Intended uses & limitations
159
 
160
+ The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model.
161
 
162
  ## Training and evaluation data
163
 
164
+ For more details about LimaRP, see the dataset page.
165
 
166
  ## Training procedure
167
 
 
179
  - lr_scheduler_warmup_steps: 10
180
  - num_epochs: 2
181
 
 
 
 
 
182
  ### Framework versions
183
 
184
  - PEFT 0.7.2.dev0