jordiclive commited on
Commit
f791de6
1 Parent(s): fac9893

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +80 -13
README.md CHANGED
@@ -34,6 +34,18 @@ model = AutoModelForCausalLM.from_pretrained(
34
  ```
35
 
36
 
 
 
 
 
 
 
 
 
 
 
 
 
37
 
38
  # LoRA Adapter for Falcon 40B trained on oasst-top1
39
 
@@ -42,7 +54,8 @@ This repo contains a **Falcon 40B** LoRA fine-tuned model and the low-rank adapt
42
 
43
  This version of the weights was trained with the following hyperparameters:
44
 
45
- - Epochs: 8
 
46
  - Batch size: 128
47
  - Max Length: 2048
48
  - Learning rate: 1e-4
@@ -50,25 +63,79 @@ This version of the weights was trained with the following hyperparameters:
50
  - Lora Alpha: 16
51
  - Lora target modules: ["dense_4h_to_h", "dense", "query_key_value", "dense_h_to_4h"]
52
 
53
- These are recommended from the QLoRA paper. The model was trained with flash attention and gradient checkpointing and deepspeed stage 3 on 8 x A100 80gb
 
 
 
 
54
 
55
 
56
  Dataset:
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  ```
58
- oasst-top1:
59
- datasets:
60
- - oasst_export:
61
- lang: "bg,ca,cs,da,de,en,es,fr,hr,hu,it,nl,pl,pt,ro,ru,sl,sr,sv,uk" # sft-8.0
62
- input_file_path: 2023-05-06_OASST_labels.jsonl.gz
63
- val_split: 0.05
64
- top_k: 1
 
65
  ```
66
 
67
- ## Model Details
68
 
69
- - **Developed** as part of the OpenAssistant Project
70
- - **Model type:** PEFT Adapter for frozen Falcon
71
- - **Language:** English
72
 
73
  ## Prompting
74
 
 
34
  ```
35
 
36
 
37
+ ## Model Details
38
+
39
+
40
+ - **Developed** as part of the OpenAssistant Project
41
+ - **Model type:** LoRA (PEFT)
42
+ - **Language:** English, German, Spanish, French (and limited capabilities in Italian, Portuguese, Polish, Dutch, Romanian, Czech, Swedish);
43
+ - **Finetuned from:** [tiiuae/falcon-40b](https://huggingface.co/tiiuae/falcon-4b)
44
+ - **Model type:** Causal decoder-only transformer language model
45
+ - **Weights & Biases:** [Training log1](https://wandb.ai/open-assistant/public-sft/runs/q0q9lce4)
46
+ [Training log2](https://wandb.ai/open-assistant/public-sft/runs/qqok9ru2?workspace=user-jordanclive)
47
+
48
+
49
 
50
  # LoRA Adapter for Falcon 40B trained on oasst-top1
51
 
 
54
 
55
  This version of the weights was trained with the following hyperparameters:
56
 
57
+ SFT 1
58
+ - Epochs: 2
59
  - Batch size: 128
60
  - Max Length: 2048
61
  - Learning rate: 1e-4
 
63
  - Lora Alpha: 16
64
  - Lora target modules: ["dense_4h_to_h", "dense", "query_key_value", "dense_h_to_4h"]
65
 
66
+ SFT2
67
+ - Epochs: 10
68
+ - Batch size: 128
69
+
70
+ The model was trained with flash attention and gradient checkpointing and deepspeed stage 3 on 8 x A100 80gb
71
 
72
 
73
  Dataset:
74
+ SFT1:
75
+ ```
76
+ - oa_leet10k:
77
+ val_split: 0.05
78
+ max_val_set: 250
79
+ - cmu_wiki_qa:
80
+ val_split: 0.05
81
+ - joke:
82
+ val_split: 0.05
83
+ - webgpt:
84
+ val_split: 0.05
85
+ max_val_set: 250
86
+ - alpaca_gpt4:
87
+ val_split: 0.025
88
+ max_val_set: 250
89
+ - gpteacher_roleplay:
90
+ val_split: 0.05
91
+ - wizardlm_70k:
92
+ val_split: 0.05
93
+ max_val_set: 500
94
+ - poem_instructions:
95
+ val_split: 0.025
96
+ - tell_a_joke:
97
+ val_split: 0.05
98
+ max_val_set: 250
99
+ - gpt4all:
100
+ val_split: 0.01
101
+ max_val_set: 1000
102
+ - minimath:
103
+ val_split: 0.05
104
+ - humaneval_mbpp_codegen_qa:
105
+ val_split: 0.05
106
+ - humaneval_mbpp_testgen_qa:
107
+ val_split: 0.05
108
+ - dolly15k:
109
+ val_split: 0.05
110
+ max_val_set: 300
111
+ - recipes:
112
+ val_split: 0.05
113
+ - code_alpaca:
114
+ val_split: 0.05
115
+ max_val_set: 250
116
+ - vicuna:
117
+ fraction: 0.5
118
+ val_split: 0.025
119
+ max_val_set: 250
120
+ - oa_wiki_qa_bart_10000row:
121
+ val_split: 0.05
122
+ max_val_set: 250
123
+ - grade_school_math_instructions:
124
+ val_split: 0.05
125
+ ```
126
+ SFT2
127
  ```
128
+ - oasst_export:
129
+ lang: "bg,ca,cs,da,de,en,es,fr,hr,hu,it,nl,pl,pt,ro,ru,sl,sr,sv,uk" # sft-8.0
130
+ input_file_path: 2023-05-06_OASST_labels.jsonl.gz
131
+ val_split: 0.05
132
+ top_k: 1
133
+ - lima:
134
+ val_split: 0.05
135
+ max_val_set: 50
136
  ```
137
 
 
138
 
 
 
 
139
 
140
  ## Prompting
141