jordiclive
commited on
Commit
•
f791de6
1
Parent(s):
fac9893
Update README.md
Browse files
README.md
CHANGED
@@ -34,6 +34,18 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
34 |
```
|
35 |
|
36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
|
38 |
# LoRA Adapter for Falcon 40B trained on oasst-top1
|
39 |
|
@@ -42,7 +54,8 @@ This repo contains a **Falcon 40B** LoRA fine-tuned model and the low-rank adapt
|
|
42 |
|
43 |
This version of the weights was trained with the following hyperparameters:
|
44 |
|
45 |
-
|
|
|
46 |
- Batch size: 128
|
47 |
- Max Length: 2048
|
48 |
- Learning rate: 1e-4
|
@@ -50,25 +63,79 @@ This version of the weights was trained with the following hyperparameters:
|
|
50 |
- Lora Alpha: 16
|
51 |
- Lora target modules: ["dense_4h_to_h", "dense", "query_key_value", "dense_h_to_4h"]
|
52 |
|
53 |
-
|
|
|
|
|
|
|
|
|
54 |
|
55 |
|
56 |
Dataset:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
57 |
```
|
58 |
-
|
59 |
-
|
60 |
-
-
|
61 |
-
|
62 |
-
|
63 |
-
|
64 |
-
|
|
|
65 |
```
|
66 |
|
67 |
-
## Model Details
|
68 |
|
69 |
-
- **Developed** as part of the OpenAssistant Project
|
70 |
-
- **Model type:** PEFT Adapter for frozen Falcon
|
71 |
-
- **Language:** English
|
72 |
|
73 |
## Prompting
|
74 |
|
|
|
34 |
```
|
35 |
|
36 |
|
37 |
+
## Model Details
|
38 |
+
|
39 |
+
|
40 |
+
- **Developed** as part of the OpenAssistant Project
|
41 |
+
- **Model type:** LoRA (PEFT)
|
42 |
+
- **Language:** English, German, Spanish, French (and limited capabilities in Italian, Portuguese, Polish, Dutch, Romanian, Czech, Swedish);
|
43 |
+
- **Finetuned from:** [tiiuae/falcon-40b](https://huggingface.co/tiiuae/falcon-4b)
|
44 |
+
- **Model type:** Causal decoder-only transformer language model
|
45 |
+
- **Weights & Biases:** [Training log1](https://wandb.ai/open-assistant/public-sft/runs/q0q9lce4)
|
46 |
+
[Training log2](https://wandb.ai/open-assistant/public-sft/runs/qqok9ru2?workspace=user-jordanclive)
|
47 |
+
|
48 |
+
|
49 |
|
50 |
# LoRA Adapter for Falcon 40B trained on oasst-top1
|
51 |
|
|
|
54 |
|
55 |
This version of the weights was trained with the following hyperparameters:
|
56 |
|
57 |
+
SFT 1
|
58 |
+
- Epochs: 2
|
59 |
- Batch size: 128
|
60 |
- Max Length: 2048
|
61 |
- Learning rate: 1e-4
|
|
|
63 |
- Lora Alpha: 16
|
64 |
- Lora target modules: ["dense_4h_to_h", "dense", "query_key_value", "dense_h_to_4h"]
|
65 |
|
66 |
+
SFT2
|
67 |
+
- Epochs: 10
|
68 |
+
- Batch size: 128
|
69 |
+
|
70 |
+
The model was trained with flash attention and gradient checkpointing and deepspeed stage 3 on 8 x A100 80gb
|
71 |
|
72 |
|
73 |
Dataset:
|
74 |
+
SFT1:
|
75 |
+
```
|
76 |
+
- oa_leet10k:
|
77 |
+
val_split: 0.05
|
78 |
+
max_val_set: 250
|
79 |
+
- cmu_wiki_qa:
|
80 |
+
val_split: 0.05
|
81 |
+
- joke:
|
82 |
+
val_split: 0.05
|
83 |
+
- webgpt:
|
84 |
+
val_split: 0.05
|
85 |
+
max_val_set: 250
|
86 |
+
- alpaca_gpt4:
|
87 |
+
val_split: 0.025
|
88 |
+
max_val_set: 250
|
89 |
+
- gpteacher_roleplay:
|
90 |
+
val_split: 0.05
|
91 |
+
- wizardlm_70k:
|
92 |
+
val_split: 0.05
|
93 |
+
max_val_set: 500
|
94 |
+
- poem_instructions:
|
95 |
+
val_split: 0.025
|
96 |
+
- tell_a_joke:
|
97 |
+
val_split: 0.05
|
98 |
+
max_val_set: 250
|
99 |
+
- gpt4all:
|
100 |
+
val_split: 0.01
|
101 |
+
max_val_set: 1000
|
102 |
+
- minimath:
|
103 |
+
val_split: 0.05
|
104 |
+
- humaneval_mbpp_codegen_qa:
|
105 |
+
val_split: 0.05
|
106 |
+
- humaneval_mbpp_testgen_qa:
|
107 |
+
val_split: 0.05
|
108 |
+
- dolly15k:
|
109 |
+
val_split: 0.05
|
110 |
+
max_val_set: 300
|
111 |
+
- recipes:
|
112 |
+
val_split: 0.05
|
113 |
+
- code_alpaca:
|
114 |
+
val_split: 0.05
|
115 |
+
max_val_set: 250
|
116 |
+
- vicuna:
|
117 |
+
fraction: 0.5
|
118 |
+
val_split: 0.025
|
119 |
+
max_val_set: 250
|
120 |
+
- oa_wiki_qa_bart_10000row:
|
121 |
+
val_split: 0.05
|
122 |
+
max_val_set: 250
|
123 |
+
- grade_school_math_instructions:
|
124 |
+
val_split: 0.05
|
125 |
+
```
|
126 |
+
SFT2
|
127 |
```
|
128 |
+
- oasst_export:
|
129 |
+
lang: "bg,ca,cs,da,de,en,es,fr,hr,hu,it,nl,pl,pt,ro,ru,sl,sr,sv,uk" # sft-8.0
|
130 |
+
input_file_path: 2023-05-06_OASST_labels.jsonl.gz
|
131 |
+
val_split: 0.05
|
132 |
+
top_k: 1
|
133 |
+
- lima:
|
134 |
+
val_split: 0.05
|
135 |
+
max_val_set: 50
|
136 |
```
|
137 |
|
|
|
138 |
|
|
|
|
|
|
|
139 |
|
140 |
## Prompting
|
141 |
|