lemonilia commited on
Commit
e653a78
1 Parent(s): 50c1efe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -10
README.md CHANGED
@@ -7,8 +7,8 @@ tags:
7
  - not-for-all-audiences
8
  ---
9
 
10
- # Limamono-7B (Mistral) v0.43
11
- This is an **early version** (43% completed) of a strongly NSFW roleplaying model trained with
12
  _extremely limited_ amounts of almost entirely synthetic data of hopefully higher quality than typical
13
  human conversations. The intended target audience is straight men and lesbians.
14
 
@@ -154,7 +154,7 @@ in the repository.
154
  ## Text generation settings
155
  For testing I use these settings:
156
  - Temperature: 1.0
157
- - Tail-Free Sampling: 0.85–0.89
158
  - Repetition Penalty: 1.11
159
  - Repetition Penalty range: 2048
160
  - Top-p: 1 (disabled), Top-k: 0 (disabled)
@@ -163,7 +163,7 @@ For testing I use these settings:
163
  [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) was used for training
164
  on one NVidia RTX3090.
165
 
166
- The training data consisted of **43** conversations (171k tokens / 963 messages)
167
  of roughly 4k tokens length. The learning rate is the one that about minimizes the
168
  eval loss on one epoch with a constant learning schedule. For the following two epochs
169
  what would be normally considered overfitting occurs, but at the same time output
@@ -182,7 +182,7 @@ quality also improves.
182
  - micro_batch_size: 1
183
  - num_epochs: 3
184
  - optimizer: adamw_torch
185
- - lr_scheduler: constant
186
  - learning_rate: 0.0002
187
  - weight_decay: 0.1
188
  - train_on_inputs: false
@@ -192,8 +192,4 @@ quality also improves.
192
  - tf32: true
193
 
194
  ### Train loss graph
195
- This one was obtained by experimentally repeating the data 3 times and finetuning for 1 epoch,
196
- with similar end results but a smoother graph without sudden jumps compared to finetuning
197
- unique data for 3 epochs.
198
-
199
- ![Train loss](https://files.catbox.moe/hiu9ah.png)
 
7
  - not-for-all-audiences
8
  ---
9
 
10
+ # Limamono-7B (Mistral) v0.50
11
+ This is an **early version** (50% completed) of a strongly NSFW roleplaying model trained with
12
  _extremely limited_ amounts of almost entirely synthetic data of hopefully higher quality than typical
13
  human conversations. The intended target audience is straight men and lesbians.
14
 
 
154
  ## Text generation settings
155
  For testing I use these settings:
156
  - Temperature: 1.0
157
+ - Tail-Free Sampling: 0.85
158
  - Repetition Penalty: 1.11
159
  - Repetition Penalty range: 2048
160
  - Top-p: 1 (disabled), Top-k: 0 (disabled)
 
163
  [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) was used for training
164
  on one NVidia RTX3090.
165
 
166
+ The training data consisted of **50** conversations (199k tokens / 1117 messages)
167
  of roughly 4k tokens length. The learning rate is the one that about minimizes the
168
  eval loss on one epoch with a constant learning schedule. For the following two epochs
169
  what would be normally considered overfitting occurs, but at the same time output
 
182
  - micro_batch_size: 1
183
  - num_epochs: 3
184
  - optimizer: adamw_torch
185
+ - lr_scheduler: cosine
186
  - learning_rate: 0.0002
187
  - weight_decay: 0.1
188
  - train_on_inputs: false
 
192
  - tf32: true
193
 
194
  ### Train loss graph
195
+ ![Train loss](https://files.catbox.moe/dg4qww.png)