Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
datasets:
|
3 |
+
- hivaze/emphatical_daily_dialogues
|
4 |
+
---
|
5 |
+
# Model Card for Model ID
|
6 |
+
|
7 |
+
This model is a adapter for databricks/dolly-v2-3b, finetuned on hivaze/emphatical_daily_dialogues.
|
8 |
+
Main goal of this model is to train model to create emphatical dialogues, which are controlled by instructions.
|
9 |
+
|
10 |
+
## Model Details
|
11 |
+
|
12 |
+
### Model Description
|
13 |
+
|
14 |
+
Prompt template: `"{intro}\n\n### Instruction:\n{instruction}\n\n### Response:\n{response}\n"`\
|
15 |
+
Example intro: "You are a kind and empathetic interlocutor. You are talking to a person. Below is an instruction that describes a task. Write a response that appropriately completes the request" \
|
16 |
+
Example instruction: "You try to chit-chat. Complete a phrase, acting like an interlocutor."
|
17 |
+
|
18 |
+
Training params:
|
19 |
+
```
|
20 |
+
train_args = TrainingArguments(
|
21 |
+
per_device_train_batch_size=8, # can be 4 with llama
|
22 |
+
per_device_eval_batch_size=8, # can be 4 with llama
|
23 |
+
gradient_accumulation_steps=4,
|
24 |
+
warmup_steps=20,
|
25 |
+
# max_steps=200,
|
26 |
+
optim="adamw_torch",
|
27 |
+
learning_rate=4e-5, # many possible values here from 1e-5 to 2e-4
|
28 |
+
# save_strategy="steps",
|
29 |
+
fp16=True,
|
30 |
+
# bf16=True, # a100 required
|
31 |
+
num_train_epochs=2,
|
32 |
+
evaluation_strategy="steps",
|
33 |
+
eval_steps=50,
|
34 |
+
save_strategy="steps",
|
35 |
+
save_steps=400,
|
36 |
+
logging_strategy="steps",
|
37 |
+
logging_steps=10,
|
38 |
+
logging_dir=f"{local_output_dir}/runs",
|
39 |
+
report_to="tensorboard",
|
40 |
+
output_dir=local_output_dir
|
41 |
+
)
|
42 |
+
```
|
43 |
+
|
44 |
+
LoRA config:
|
45 |
+
```
|
46 |
+
config = LoraConfig(
|
47 |
+
r=16, # can be 8 with llama
|
48 |
+
lora_alpha=32, # can be 16 with llama
|
49 |
+
# target_modules=["q_proj", "v_proj"],
|
50 |
+
target_modules=['query_key_value'],
|
51 |
+
lora_dropout=0.05,
|
52 |
+
bias="none",
|
53 |
+
task_type="CAUSAL_LM"
|
54 |
+
)
|
55 |
+
```
|
56 |
+
|
57 |
+
- **Developed by:** hivaze
|
58 |
+
- **Model type:** LoRA adapter for GPTNeoXForCausalLM
|
59 |
+
- **Language(s) (NLP):** Primarly english
|
60 |
+
- **Finetuned from model [optional]:** databricks/dolly-v2-3b
|
61 |
+
- **Git repository**: https://github.com/hivaze/friendly_chatbot_task
|
62 |
+
|
63 |
+
### Tensorboard
|
64 |
+
|
65 |
+
![image.png](https://s3.amazonaws.com/moonup/production/uploads/648e72a866dcba8b5aaecbdc/ldy8tEvYPlTCo-2T1WElf.png)
|
66 |
+
|
67 |
+
|