jpark677 commited on
Commit
049a871
·
verified ·
1 Parent(s): 2f6afbf

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.4.0
adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "liuhaotian/llava-v1.5-7b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 256,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 128,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "k_proj",
18
+ "up_proj",
19
+ "down_proj",
20
+ "gate_proj",
21
+ "v_proj",
22
+ "q_proj",
23
+ "o_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a8793dda23484b317fbe9c97b5add4bd1c48e4b80509149a9badab1a2fdcd6e
3
+ size 639787082
config.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "liuhaotian/llava-v1.5-7b",
3
+ "architectures": [
4
+ "LlavaLlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "freeze_mm_mlp_adapter": false,
11
+ "freeze_mm_vision_resampler": false,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "image_aspect_ratio": "pad",
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 11008,
17
+ "max_length": 4096,
18
+ "max_position_embeddings": 4096,
19
+ "mm_hidden_size": 1024,
20
+ "mm_patch_merge_type": "flat",
21
+ "mm_projector_lr": 2e-05,
22
+ "mm_projector_type": "mlp2x_gelu",
23
+ "mm_resampler_type": null,
24
+ "mm_use_im_patch_token": false,
25
+ "mm_use_im_start_end": false,
26
+ "mm_vision_select_feature": "patch",
27
+ "mm_vision_select_layer": -2,
28
+ "mm_vision_tower": "openai/clip-vit-large-patch14-336",
29
+ "model_type": "llava_llama",
30
+ "num_attention_heads": 32,
31
+ "num_hidden_layers": 32,
32
+ "num_key_value_heads": 32,
33
+ "pad_token_id": 0,
34
+ "pretraining_tp": 1,
35
+ "rms_norm_eps": 1e-05,
36
+ "rope_scaling": null,
37
+ "rope_theta": 10000.0,
38
+ "tie_word_embeddings": false,
39
+ "tokenizer_model_max_length": 2048,
40
+ "tokenizer_padding_side": "right",
41
+ "torch_dtype": "float16",
42
+ "transformers_version": "4.37.2",
43
+ "tune_mm_mlp_adapter": false,
44
+ "tune_mm_vision_resampler": false,
45
+ "unfreeze_mm_vision_tower": false,
46
+ "use_cache": true,
47
+ "use_mm_proj": true,
48
+ "vocab_size": 32000
49
+ }
non_lora_trainables.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2ab232df12ada0289917555966e519bd3f14e71538eaa68b7d174010f3e8b6e
3
+ size 41961648
trainer_state.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 35,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.03,
13
+ "learning_rate": 0.0001,
14
+ "loss": 1.41,
15
+ "step": 1
16
+ },
17
+ {
18
+ "epoch": 0.06,
19
+ "learning_rate": 0.0002,
20
+ "loss": 1.3381,
21
+ "step": 2
22
+ },
23
+ {
24
+ "epoch": 0.09,
25
+ "learning_rate": 0.00019954719225730847,
26
+ "loss": 0.3192,
27
+ "step": 3
28
+ },
29
+ {
30
+ "epoch": 0.11,
31
+ "learning_rate": 0.00019819286972627066,
32
+ "loss": 0.333,
33
+ "step": 4
34
+ },
35
+ {
36
+ "epoch": 0.14,
37
+ "learning_rate": 0.00019594929736144976,
38
+ "loss": 0.3892,
39
+ "step": 5
40
+ },
41
+ {
42
+ "epoch": 0.17,
43
+ "learning_rate": 0.00019283679330160726,
44
+ "loss": 0.2928,
45
+ "step": 6
46
+ },
47
+ {
48
+ "epoch": 0.2,
49
+ "learning_rate": 0.00018888354486549237,
50
+ "loss": 0.3289,
51
+ "step": 7
52
+ },
53
+ {
54
+ "epoch": 0.23,
55
+ "learning_rate": 0.00018412535328311814,
56
+ "loss": 0.2921,
57
+ "step": 8
58
+ },
59
+ {
60
+ "epoch": 0.26,
61
+ "learning_rate": 0.00017860530947427875,
62
+ "loss": 0.2556,
63
+ "step": 9
64
+ },
65
+ {
66
+ "epoch": 0.29,
67
+ "learning_rate": 0.00017237340381050703,
68
+ "loss": 0.3128,
69
+ "step": 10
70
+ },
71
+ {
72
+ "epoch": 0.31,
73
+ "learning_rate": 0.00016548607339452853,
74
+ "loss": 0.2491,
75
+ "step": 11
76
+ },
77
+ {
78
+ "epoch": 0.34,
79
+ "learning_rate": 0.00015800569095711982,
80
+ "loss": 0.2943,
81
+ "step": 12
82
+ },
83
+ {
84
+ "epoch": 0.37,
85
+ "learning_rate": 0.00015000000000000001,
86
+ "loss": 0.2836,
87
+ "step": 13
88
+ },
89
+ {
90
+ "epoch": 0.4,
91
+ "learning_rate": 0.00014154150130018866,
92
+ "loss": 0.2055,
93
+ "step": 14
94
+ },
95
+ {
96
+ "epoch": 0.43,
97
+ "learning_rate": 0.00013270679633174218,
98
+ "loss": 0.1871,
99
+ "step": 15
100
+ },
101
+ {
102
+ "epoch": 0.46,
103
+ "learning_rate": 0.00012357589355094275,
104
+ "loss": 0.2089,
105
+ "step": 16
106
+ },
107
+ {
108
+ "epoch": 0.49,
109
+ "learning_rate": 0.00011423148382732853,
110
+ "loss": 0.2982,
111
+ "step": 17
112
+ },
113
+ {
114
+ "epoch": 0.51,
115
+ "learning_rate": 0.00010475819158237425,
116
+ "loss": 0.1497,
117
+ "step": 18
118
+ },
119
+ {
120
+ "epoch": 0.54,
121
+ "learning_rate": 9.524180841762577e-05,
122
+ "loss": 0.2389,
123
+ "step": 19
124
+ },
125
+ {
126
+ "epoch": 0.57,
127
+ "learning_rate": 8.57685161726715e-05,
128
+ "loss": 0.2047,
129
+ "step": 20
130
+ },
131
+ {
132
+ "epoch": 0.6,
133
+ "learning_rate": 7.642410644905726e-05,
134
+ "loss": 0.1506,
135
+ "step": 21
136
+ },
137
+ {
138
+ "epoch": 0.63,
139
+ "learning_rate": 6.729320366825784e-05,
140
+ "loss": 0.1873,
141
+ "step": 22
142
+ },
143
+ {
144
+ "epoch": 0.66,
145
+ "learning_rate": 5.845849869981137e-05,
146
+ "loss": 0.1362,
147
+ "step": 23
148
+ },
149
+ {
150
+ "epoch": 0.69,
151
+ "learning_rate": 5.000000000000002e-05,
152
+ "loss": 0.1933,
153
+ "step": 24
154
+ },
155
+ {
156
+ "epoch": 0.71,
157
+ "learning_rate": 4.19943090428802e-05,
158
+ "loss": 0.2071,
159
+ "step": 25
160
+ },
161
+ {
162
+ "epoch": 0.74,
163
+ "learning_rate": 3.45139266054715e-05,
164
+ "loss": 0.1955,
165
+ "step": 26
166
+ },
167
+ {
168
+ "epoch": 0.77,
169
+ "learning_rate": 2.7626596189492983e-05,
170
+ "loss": 0.1245,
171
+ "step": 27
172
+ },
173
+ {
174
+ "epoch": 0.8,
175
+ "learning_rate": 2.139469052572127e-05,
176
+ "loss": 0.1603,
177
+ "step": 28
178
+ },
179
+ {
180
+ "epoch": 0.83,
181
+ "learning_rate": 1.587464671688187e-05,
182
+ "loss": 0.1888,
183
+ "step": 29
184
+ },
185
+ {
186
+ "epoch": 0.86,
187
+ "learning_rate": 1.1116455134507664e-05,
188
+ "loss": 0.17,
189
+ "step": 30
190
+ },
191
+ {
192
+ "epoch": 0.89,
193
+ "learning_rate": 7.163206698392744e-06,
194
+ "loss": 0.1368,
195
+ "step": 31
196
+ },
197
+ {
198
+ "epoch": 0.91,
199
+ "learning_rate": 4.050702638550275e-06,
200
+ "loss": 0.1649,
201
+ "step": 32
202
+ },
203
+ {
204
+ "epoch": 0.94,
205
+ "learning_rate": 1.8071302737293295e-06,
206
+ "loss": 0.1879,
207
+ "step": 33
208
+ },
209
+ {
210
+ "epoch": 0.97,
211
+ "learning_rate": 4.5280774269154115e-07,
212
+ "loss": 0.1047,
213
+ "step": 34
214
+ },
215
+ {
216
+ "epoch": 1.0,
217
+ "learning_rate": 0.0,
218
+ "loss": 0.1412,
219
+ "step": 35
220
+ },
221
+ {
222
+ "epoch": 1.0,
223
+ "step": 35,
224
+ "total_flos": 4772403118080.0,
225
+ "train_loss": 0.28688187726906367,
226
+ "train_runtime": 248.2823,
227
+ "train_samples_per_second": 17.629,
228
+ "train_steps_per_second": 0.141
229
+ }
230
+ ],
231
+ "logging_steps": 1.0,
232
+ "max_steps": 35,
233
+ "num_input_tokens_seen": 0,
234
+ "num_train_epochs": 1,
235
+ "save_steps": 50000,
236
+ "total_flos": 4772403118080.0,
237
+ "train_batch_size": 16,
238
+ "trial_name": null,
239
+ "trial_params": null
240
+ }