Upload README.md
Browse files
README.md
CHANGED
@@ -183,19 +183,10 @@ model = AutoModelForCausalLM.from_pretrained(model_name_or_path,
|
|
183 |
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
184 |
|
185 |
prompt = "Tell me about AI"
|
186 |
-
prompt_template=f'''
|
187 |
-
|
188 |
-
The `<|system|>` prompt can be used to inject out-of-channel information behind the scenes, while the `<|user|>` prompt should be used to indicate user input.
|
189 |
-
The `<|model|>` token should then be used to indicate that the model should generate a response. These tokens can happen multiple times and be chained up to form a conversation history.
|
190 |
-
|
191 |
-
The system prompt has been designed to allow the model to "enter" various modes and dictate the reply length. Here's an example:
|
192 |
-
|
193 |
-
```
|
194 |
-
<|system|>Enter RP mode. Pretend to be {{char}} whose persona follows:
|
195 |
{{persona}}
|
196 |
|
197 |
You shall reply to the user while staying in character, and generate long responses.
|
198 |
-
```
|
199 |
|
200 |
'''
|
201 |
|
|
|
183 |
tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)
|
184 |
|
185 |
prompt = "Tell me about AI"
|
186 |
+
prompt_template=f'''<|system|>Enter RP mode. Pretend to be {{char}} whose persona follows:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
187 |
{{persona}}
|
188 |
|
189 |
You shall reply to the user while staying in character, and generate long responses.
|
|
|
190 |
|
191 |
'''
|
192 |
|