922CA commited on
Commit
85fb924
1 Parent(s): e151918

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -6
README.md CHANGED
@@ -6,16 +6,29 @@ license: llama2
6
  * Llama 7b chat fine-tuned for Monika character from DDLC
7
  * Trained on a dataset of ~600 items (dialogue scraped from game, reddit, and Twitter augmented by Nous Hermes 13b to turn each into snippets of multi-turn chat dialogue between Player and Monika + manually crafted test dataset of 12 items)
8
  * Earlier variant trained without dataset augmentation [here](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2e)
 
 
 
 
 
 
 
 
 
 
 
9
  * Trained for 2 epochs
 
 
 
 
 
 
 
10
 
11
  ### WARNINGS AND DISCLAIMERS
12
  Note that aside from formatting and other minor edits, dataset used is mostly as is generated by LM. As such, while this version is better at coherency or chatting than previous ones, it may not reflect perfectly the characteristics of Monika (i.e. she will claim to have an office, work as a translator, or play the guitar). Next version will train on a manually curated and edited version of this dataset, where dialogue will be edited to reflect her characteristics more.
13
 
14
  Also looking to switch to a different a base model to work off from for future versions aside from llama 7b chat.
15
 
16
- Finally, this model is not guaranteed to output aligned or safe outputs, use at your own risk.
17
-
18
- ### USAGE
19
- For best results: replace "Human" and "Assistant" with "Player" and "Monika" like so:
20
-
21
- \nPlayer: (prompt)\nMonika:
 
6
  * Llama 7b chat fine-tuned for Monika character from DDLC
7
  * Trained on a dataset of ~600 items (dialogue scraped from game, reddit, and Twitter augmented by Nous Hermes 13b to turn each into snippets of multi-turn chat dialogue between Player and Monika + manually crafted test dataset of 12 items)
8
  * Earlier variant trained without dataset augmentation [here](https://huggingface.co/922-CA/llama-2-7b-delphi-v0.2e)
9
+ * [GGMLs](https://huggingface.co/922-CA/llama-2-7b-monika-v0.3b-ggml)
10
+ * [QLoras (hf and GGML)](https://huggingface.co/922-CA/monika-lm-lora-tests/tree/main/llama-2-7b-chat-monika-v0.3b)
11
+
12
+ ### USAGE
13
+ This is meant to be mainly a chat model with limited RP ability.
14
+
15
+ For best results: replace "Human" and "Assistant" with "Player" and "Monika" like so:
16
+
17
+ \nPlayer: (prompt)\nMonika:
18
+
19
+ ### HYPERPARAMS
20
  * Trained for 2 epochs
21
+ * rank: 64
22
+ * lora alpha: 16
23
+ * lora dropout: 0.1
24
+ * lr: 2e-4
25
+ * batch size: 8
26
+ * warmup ratio: 0.03
27
+ * grad steps: 1
28
 
29
  ### WARNINGS AND DISCLAIMERS
30
  Note that aside from formatting and other minor edits, dataset used is mostly as is generated by LM. As such, while this version is better at coherency or chatting than previous ones, it may not reflect perfectly the characteristics of Monika (i.e. she will claim to have an office, work as a translator, or play the guitar). Next version will train on a manually curated and edited version of this dataset, where dialogue will be edited to reflect her characteristics more.
31
 
32
  Also looking to switch to a different a base model to work off from for future versions aside from llama 7b chat.
33
 
34
+ Finally, this model is not guaranteed to output aligned or safe outputs, use at your own risk.