nxnhjrjtbjfzhrovwl commited on
Commit
c89118a
1 Parent(s): a76ef12

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -11
README.md CHANGED
@@ -4,7 +4,7 @@
4
  {{ card_data }}
5
  ---
6
 
7
- This repository contains the unquantized merge of [limarp-llama2 7b lora](https://huggingface.co/lemonilia/limarp-llama2) in ggml format.
8
 
9
  The below is the contents of the original model card:
10
 
@@ -12,6 +12,8 @@ The below is the contents of the original model card:
12
 
13
  LIMARP-Llama2 is an experimental [Llama2](https://huggingface.co/meta-llama) finetune narrowly focused on novel-style roleplay chatting.
14
 
 
 
15
  ## Model Details
16
 
17
  ### Model Description
@@ -56,7 +58,13 @@ It's possible to make the model automatically generate random character informat
56
 
57
  ![example](https://files.catbox.moe/5ntmcj.png)
58
 
59
- Here is an example SillyTavern character card following the intended format: https://files.catbox.moe/r20w0r.png (download and import into SillyTavern)
 
 
 
 
 
 
60
 
61
  And here is a sample of how the model is intended to behave with proper chat and prompt formatting: https://files.catbox.moe/egfd90.png
62
 
@@ -73,7 +81,7 @@ And here is a sample of how the model is intended to behave with proper chat and
73
  - Suggested text generation settings:
74
  - Temperature ~0.70
75
  - Tail-Free Sampling 0.85
76
- - Repetition penalty 1.05~1.10 (lower values preferred for stability)
77
  - Not used: Top-P (disabled/set to 1.0), Top-K (disabled/set to 0), Typical P (disabled/set to 1.0)
78
 
79
 
@@ -84,9 +92,9 @@ And here is a sample of how the model is intended to behave with proper chat and
84
  The model has not been tested for:
85
 
86
  - IRC-style chat
87
- - Markdown-style roleplay (asterisks for actions)
88
  - Storywriting
89
- - Usage without the suggested prompt format (it will output short and boring responses without it)
90
 
91
  Furthermore, the model is not intended nor expected to provide factual and accurate information on any subject.
92
 
@@ -100,17 +108,16 @@ The model will show biases similar to those observed in niche roleplaying forums
100
 
101
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
102
 
103
- The model may easily output disturbing and socially inappropriate content and therefore should not be used by minors or within environments where a general audience is expected.
104
 
105
  ## How to Get Started with the Model
106
 
107
- Download and load with `text-generation-webui` as a back-end application. It's suggested to start the `webui` via command line. Assuming you have copied the LoRA files under a subdirectory called `lora/limarp-llama2`, you would use something like this:
108
 
109
  ```
110
  python server.py --api --verbose --model Llama-7B --lora limarp-llama2
111
  ```
112
 
113
-
114
  Then, preferably use [SillyTavern](https://github.com/SillyTavern/SillyTavern) as a front-end using the following settings:
115
 
116
  ![SillyTavern settings](https://i.imgur.com/gDPC8gx.png)
@@ -129,7 +136,7 @@ To take advantage of this model's larger context length, unlock the context size
129
 
130
  <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
131
 
132
- The training data comprises **1005** manually edited roleplaying conversation threads from various Internet RP forums, for about 11 megabytes of training data.
133
 
134
  Character and Scenario information was filled in for every thread with the help of mainly `gpt-4`, but otherwise conversations in the dataset are almost entirely human-generated except for a handful of messages. Character names in the RP stories have been isolated and replaced with standard placeholder strings. Usernames, out-of-context (OOC) messages and personal information have not been intentionally included.
135
 
@@ -147,7 +154,7 @@ The most important settings for QLoRA were as follows:
147
  - --train_on_source True
148
  - --learning_rate 0.00006
149
  - --lr_scheduler_type cosine
150
- - --lora_r 32
151
  - --max_steps -1
152
  - --num_train_epochs 2
153
  - --bf16 True
@@ -168,4 +175,4 @@ It was also found that using `--train_on_source False` with the entire training
168
 
169
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
170
 
171
- Finetuning this model requires about 1 kWh of electricity for 2 epochs.
 
4
  {{ card_data }}
5
  ---
6
 
7
+ This repository contains the unquantized merge of [limarp-llama2 lora](https://huggingface.co/lemonilia/limarp-llama2) in ggml format.
8
 
9
  The below is the contents of the original model card:
10
 
 
12
 
13
  LIMARP-Llama2 is an experimental [Llama2](https://huggingface.co/meta-llama) finetune narrowly focused on novel-style roleplay chatting.
14
 
15
+ To considerably facilitate uploading and distribution, LoRA adapters have been provided instead of the merged models. You should get the Llama2 base model first, either from Meta or from one of the reuploads on HuggingFace (for example [here](https://huggingface.co/NousResearch/Llama-2-7b-hf) and [here](https://huggingface.co/NousResearch/Llama-2-13b-hf)). It is also possible to apply the LoRAs on different Llama2-based models (e.g. [LLongMA-2](https://huggingface.co/conceptofmind/LLongMA-2-7b) or [Nous-Hermes-Llama2](https://huggingface.co/NousResearch/Nous-Hermes-Llama2-13b)), although this is largely untested and the final results may not work as intended.
16
+
17
  ## Model Details
18
 
19
  ### Model Description
 
58
 
59
  ![example](https://files.catbox.moe/5ntmcj.png)
60
 
61
+ Here are a few example **SillyTavern character cards** following the required format; download and import into SillyTavern. Feel free to modify and adapt them to your purposes.
62
+
63
+ - [Carina, a 'big sister' android maid](https://files.catbox.moe/1qcqqj.png)
64
+ - [Charlotte, a cute android maid](https://files.catbox.moe/k1x9a7.png)
65
+ - [Etma, an 'aligned' AI assistant](https://files.catbox.moe/dj8ggi.png)
66
+ - [Mila, an anthro pet catgirl](https://files.catbox.moe/amnsew.png)
67
+ - [Samuel, a handsome vampire](https://files.catbox.moe/f9uiw1.png)
68
 
69
  And here is a sample of how the model is intended to behave with proper chat and prompt formatting: https://files.catbox.moe/egfd90.png
70
 
 
81
  - Suggested text generation settings:
82
  - Temperature ~0.70
83
  - Tail-Free Sampling 0.85
84
+ - Repetition penalty ~1.10 (Compared to LLaMAv1, Llama2 appears to require a somewhat higher rep.pen.)
85
  - Not used: Top-P (disabled/set to 1.0), Top-K (disabled/set to 0), Typical P (disabled/set to 1.0)
86
 
87
 
 
92
  The model has not been tested for:
93
 
94
  - IRC-style chat
95
+ - Markdown-style roleplay (asterisks for actions, dialogue lines without quotation marks)
96
  - Storywriting
97
+ - Usage without the suggested prompt format
98
 
99
  Furthermore, the model is not intended nor expected to provide factual and accurate information on any subject.
100
 
 
108
 
109
  <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
110
 
111
+ The model may easily output disturbing and socially inappropriate content and therefore should not be used by minors or within environments where a general audience is expected. Its outputs will have in general a strong NSFW bias unless the character card/description de-emphasizes it.
112
 
113
  ## How to Get Started with the Model
114
 
115
+ Download and load with `text-generation-webui` as a back-end application. It's suggested to start the `webui` via command line. Assuming you have copied the LoRA files under a subdirectory called `lora/limarp-llama2`, you would use something like this for the 7B model:
116
 
117
  ```
118
  python server.py --api --verbose --model Llama-7B --lora limarp-llama2
119
  ```
120
 
 
121
  Then, preferably use [SillyTavern](https://github.com/SillyTavern/SillyTavern) as a front-end using the following settings:
122
 
123
  ![SillyTavern settings](https://i.imgur.com/gDPC8gx.png)
 
136
 
137
  <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
138
 
139
+ The training data comprises **1005** manually edited roleplaying conversation threads from various Internet RP forums, for about 11 megabytes of data.
140
 
141
  Character and Scenario information was filled in for every thread with the help of mainly `gpt-4`, but otherwise conversations in the dataset are almost entirely human-generated except for a handful of messages. Character names in the RP stories have been isolated and replaced with standard placeholder strings. Usernames, out-of-context (OOC) messages and personal information have not been intentionally included.
142
 
 
154
  - --train_on_source True
155
  - --learning_rate 0.00006
156
  - --lr_scheduler_type cosine
157
+ - --lora_r 32 (7B LoRA), 8 (13B LoRA)
158
  - --max_steps -1
159
  - --num_train_epochs 2
160
  - --bf16 True
 
175
 
176
  <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
177
 
178
+ Finetuning this model requires about 1 kWh (7B LoRA) or 2.1 kWh (13B LoRA) of electricity for 2 epochs, excluding testing.