Update README.md
Browse files
README.md
CHANGED
@@ -7,8 +7,6 @@ tags:
|
|
7 |
|
8 |
**NOTE**: See [creative-writing-control-vectors-v2.1](https://huggingface.co/jukofyork/creative-writing-control-vectors-v2.1) for the current main control-vector repo.
|
9 |
|
10 |
-
*(I will also add `WizardLM-2-8x22B`, `c4ai-command-r-v01` and `gemma-2-27b-it` versions of these control vectors in the next few days...)*
|
11 |
-
|
12 |
## Details
|
13 |
|
14 |
The control-vectors in this repo were created as an experiment by increasing the triplets in `system_messages_outlook_extended.json` by 4x (click to expand):
|
@@ -160,16 +158,16 @@ I also include 3 different values for the `--regularisation_factor` option; `1.0
|
|
160 |
|
161 |
Try to use the largest `regularisation_factor` that has the desired effect - this has the least chance of damaging the models' outputs.
|
162 |
|
163 |
-
## Prompting format for `Mistral-Large-Instruct-2407` and `
|
164 |
|
165 |
-
I have found by testing that
|
166 |
|
167 |
```
|
168 |
USER: {prompt}
|
169 |
ASSISTANT:
|
170 |
```
|
171 |
|
172 |
-
so I altered the 'Jinja2' `chat_template` in the `tokenizer_config.json` for
|
173 |
|
174 |
```json
|
175 |
{
|
@@ -177,4 +175,4 @@ so I altered the 'Jinja2' `chat_template` in the `tokenizer_config.json` for bot
|
|
177 |
}
|
178 |
```
|
179 |
|
180 |
-
|
|
|
7 |
|
8 |
**NOTE**: See [creative-writing-control-vectors-v2.1](https://huggingface.co/jukofyork/creative-writing-control-vectors-v2.1) for the current main control-vector repo.
|
9 |
|
|
|
|
|
10 |
## Details
|
11 |
|
12 |
The control-vectors in this repo were created as an experiment by increasing the triplets in `system_messages_outlook_extended.json` by 4x (click to expand):
|
|
|
158 |
|
159 |
Try to use the largest `regularisation_factor` that has the desired effect - this has the least chance of damaging the models' outputs.
|
160 |
|
161 |
+
## Prompting format for `Mistral-Large-Instruct-2407`, `WizardLM-2-8x22B` and `miqu-1-70b`:
|
162 |
|
163 |
+
I have found by testing that these models seems to work much better for creative writing if you use the following 'Vicuna' prompt template:
|
164 |
|
165 |
```
|
166 |
USER: {prompt}
|
167 |
ASSISTANT:
|
168 |
```
|
169 |
|
170 |
+
so I altered the 'Jinja2' `chat_template` in the `tokenizer_config.json` for `Mistral-Large-Instruct-2407`, `WizardLM-2-8x22B` and `miqu-1-70b` to this for the training of these control vectors:
|
171 |
|
172 |
```json
|
173 |
{
|
|
|
175 |
}
|
176 |
```
|
177 |
|
178 |
+
**NOTE**: I still used the default prompt templates for the other 3 models (`c4ai-command-r-plus`, `c4ai-command-r-v01` and `gemma-2-27b-it`).
|