Update README.md
Browse files
README.md
CHANGED
@@ -19,4 +19,23 @@ tags:
|
|
19 |
|
20 |
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
|
21 |
|
22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
|
20 |
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
|
21 |
|
22 |
+
The model is trained on reddit_depression_dataset, The epoch = 1.
|
23 |
+
|
24 |
+
The training is in dialog format, but the user's input is ignored.
|
25 |
+
|
26 |
+
For example
|
27 |
+
|
28 |
+
```python
|
29 |
+
def formatting_prompts_func(examples):
|
30 |
+
texts_dataset = examples['text']
|
31 |
+
formatted_prompts = []
|
32 |
+
for text in texts_dataset:
|
33 |
+
dialog = [
|
34 |
+
{"role": "system", "content": "You are a patient undergoing depression."},
|
35 |
+
# {"role": "user", "content": ""},
|
36 |
+
{"role": "assistant", "content": text}
|
37 |
+
]
|
38 |
+
formatted_prompt = tokenizer.apply_chat_template(dialog, tokenize=False, add_generation_prompt=False)
|
39 |
+
formatted_prompts.append(formatted_prompt)
|
40 |
+
return {"text": formatted_prompts}
|
41 |
+
```
|