Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
license_name: tongyi-qianwen
|
4 |
+
license_link: https://huggingface.co/Qwen/Qwen2-72B-Instruct/blob/main/LICENSE
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
- zh
|
8 |
+
pipeline_tag: text-generation
|
9 |
+
tags:
|
10 |
+
- chat
|
11 |
+
---
|
12 |
+
|
13 |
+
![](https://files.catbox.moe/ngqnb1.png)
|
14 |
+
|
15 |
+
This is the first in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of [Qwen-2 72B Instruct](https://huggingface.co/Qwen/Qwen2-72B-Instruct).
|
16 |
+
|
17 |
+
|
18 |
+
## Prompting
|
19 |
+
Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:
|
20 |
+
|
21 |
+
```py
|
22 |
+
"""<|im_start|>user
|
23 |
+
Hi there!<|im_end|>
|
24 |
+
<|im_start|>assistant
|
25 |
+
Nice to meet you!<|im_end|>
|
26 |
+
<|im_start|>user
|
27 |
+
Can I ask a question?<|im_end|>
|
28 |
+
<|im_start|>assistant
|
29 |
+
"""
|
30 |
+
```
|
31 |
+
|
32 |
+
## Credits
|
33 |
+
|
34 |
+
This model has been a team effort, credits go to:
|
35 |
+
|
36 |
+
- [Sao10K](https://huggingface.co/Sao10K) for help with (and cleaning up!) the dataset.
|
37 |
+
- [alpindale](https://huggingface.co/alpindale) for the training.
|
38 |
+
- [kalomaze](https://huggingface.co/kalomaze) for helping with the hyperparameter tuning.
|
39 |
+
- Various other people for their continued help as we tuned the parameters, restarted failed runs. In no particular order: [Doctor Shotgun](https://huggingface.co/Doctor-Shotgun), [Lucy](https://huggingface.co/lucyknada), [Nopm](https://huggingface.co/nopm), [Mango](https://huggingface.co/MangoMango69420), and the rest of the Silly Tilly.
|
40 |
+
|
41 |
+
And last but not least, we'd like to thank [Kearm](https://twitter.com/Nottlespike) for sponsoring the compute needed to train this model.
|
42 |
+
|
43 |
+
## Training
|
44 |
+
The training was done with 55 million tokens of high-quality RP data, over 1.5 epochs. We used 8x [AMD Instinct™ MI300X Accelerators](https://www.amd.com/en/products/accelerators/instinct/mi300/mi300x.html) for the full-parameter fine-tuning of the model.
|
45 |
+
|
46 |
+
|
47 |
+
## Safety
|
48 |
+
...
|