Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -26,8 +26,19 @@ This repo contains the model checkpoints for:
|
|
26 |
- optimized with the loss <b>PPO</b>
|
27 |
- aligned using the SHP, Anthropic HH and Open Assistant datasets.
|
28 |
|
29 |
-
To prompt
|
30 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
|
32 |
Please refer to our [code repository](https://github.com/ContextualAI/HALOs) or [blog](https://contextual.ai/better-cheaper-faster-llm-alignment-with-kto/) which contains intructions for training your own HALOs and links to our model cards.
|
33 |
|
|
|
26 |
- optimized with the loss <b>PPO</b>
|
27 |
- aligned using the SHP, Anthropic HH and Open Assistant datasets.
|
28 |
|
29 |
+
To prompt Archangel models, ensure that the format is consistent with that of TuluV2.
|
30 |
+
For example, a prompt should be formatted as follows, where `<|user|>` corresponds to the human's role and `<|assistant|>` corresponds to the LLM's role.
|
31 |
+
The human should speak first:
|
32 |
+
```
|
33 |
+
<|user|>
|
34 |
+
Hi! I'm looking for a cake recipe.
|
35 |
+
<|assistant|>
|
36 |
+
What kind of cake?
|
37 |
+
<|user|>
|
38 |
+
Chocolate cake.
|
39 |
+
<|assistant|>
|
40 |
+
```
|
41 |
+
Note that a beginning-of-sequence (BOS) token is automatically added by all Archangel models during tokenization and does not have to be added by you. No end-of-sequence (EOS) token is added to the prompt.
|
42 |
|
43 |
Please refer to our [code repository](https://github.com/ContextualAI/HALOs) or [blog](https://contextual.ai/better-cheaper-faster-llm-alignment-with-kto/) which contains intructions for training your own HALOs and links to our model cards.
|
44 |
|