hivaze commited on
Commit
929c7ac
1 Parent(s): 1034d32

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -0
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - hivaze/emphatical_daily_dialogues
4
+ ---
5
+ # Model Card for Model ID
6
+
7
+ This model is a adapter for databricks/dolly-v2-7b, finetuned on hivaze/emphatical_daily_dialogues.
8
+ Main goal of this model is to train model to create emphatical dialogues, which are controlled by instructions.
9
+
10
+ ## Model Details
11
+
12
+ ### Model Description
13
+
14
+ Prompt template: `"{intro}\n\n### Instruction:\n{instruction}\n\n### Response:\n{response}\n"`\
15
+ Example intro: "You are a kind and empathetic interlocutor. You are talking to a person. Below is an instruction that describes a task. Write a response that appropriately completes the request" \
16
+ Example instruction: "You try to chit-chat. Complete a phrase, acting like an interlocutor."
17
+
18
+ Training params:
19
+ ```
20
+ train_args = TrainingArguments(
21
+ per_device_train_batch_size=8, # can be 4 with llama
22
+ per_device_eval_batch_size=8, # can be 4 with llama
23
+ gradient_accumulation_steps=4,
24
+ warmup_steps=20,
25
+ # max_steps=200,
26
+ optim="adamw_torch",
27
+ learning_rate=4e-5, # many possible values here from 1e-5 to 2e-4
28
+ # save_strategy="steps",
29
+ fp16=True,
30
+ # bf16=True, # a100 required
31
+ num_train_epochs=1,
32
+ evaluation_strategy="steps",
33
+ eval_steps=50,
34
+ save_strategy="steps",
35
+ save_steps=400,
36
+ logging_strategy="steps",
37
+ logging_steps=10,
38
+ logging_dir=f"{local_output_dir}/runs",
39
+ report_to="tensorboard",
40
+ output_dir=local_output_dir
41
+ )
42
+ ```
43
+
44
+ LoRA config:
45
+ ```
46
+ config = LoraConfig(
47
+ r=16, # can be 8 with llama
48
+ lora_alpha=32, # can be 16 with llama
49
+ # target_modules=["q_proj", "v_proj"],
50
+ target_modules=['query_key_value'],
51
+ lora_dropout=0.05,
52
+ bias="none",
53
+ task_type="CAUSAL_LM"
54
+ )
55
+ ```
56
+
57
+ - **Developed by:** hivaze
58
+ - **Model type:** LoRA adapter for GPTNeoXForCausalLM
59
+ - **Language(s) (NLP):** Primarly english
60
+ - **Finetuned from model [optional]:** databricks/dolly-v2-7b
61
+ - **Git repository**: https://github.com/hivaze/friendly_chatbot_task
62
+
63
+ ### Tensorboard
64
+
65
+ ![image.png](https://s3.amazonaws.com/moonup/production/uploads/648e72a866dcba8b5aaecbdc/DOwmUbIW3tQJvxtRASRwz.png)
66
+
67
+