Remove image_size
Browse files- preprocessor_config.json +49 -51
preprocessor_config.json
CHANGED
@@ -1,53 +1,51 @@
|
|
1 |
{
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
41 |
-
|
42 |
-
|
43 |
-
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
|
48 |
-
|
49 |
-
|
50 |
-
|
51 |
-
"vit_remove_last":false,
|
52 |
-
"vocab_size":50265
|
53 |
}
|
|
|
1 |
{
|
2 |
+
"downstream_fusion":false,
|
3 |
+
"downstream_fusion_layers":1,
|
4 |
+
"downstream_fusion_method":"elmo",
|
5 |
+
"drop_rate":0.1,
|
6 |
+
"freeze_RoBERTa":false,
|
7 |
+
"freeze_ViT":false,
|
8 |
+
"freeze_layer_count_roberta":false,
|
9 |
+
"freeze_layer_count_vit":false,
|
10 |
+
"head_hidden_scale":2,
|
11 |
+
"hidden_size":768,
|
12 |
+
"input_text_embed_size":768,
|
13 |
+
"link_tower_shared":false,
|
14 |
+
"link_tower_type":"add",
|
15 |
+
"log_dir":"log_dir",
|
16 |
+
"loss_names":{"contras": 0,
|
17 |
+
"irtr": 0,
|
18 |
+
"itm": 0,
|
19 |
+
"mlm": 0,
|
20 |
+
"mpp": 0,
|
21 |
+
"nlvr2": 0,
|
22 |
+
"snli": 0,
|
23 |
+
"vcr": 0,
|
24 |
+
"vcr_qar": 0,
|
25 |
+
"vqa": 1},
|
26 |
+
"max_text_len":50,
|
27 |
+
"mlp_ratio":4,
|
28 |
+
"model_type":"bridgetower",
|
29 |
+
"num_heads":12,
|
30 |
+
"num_layers":6,
|
31 |
+
"num_nodes":1,
|
32 |
+
"only_load_cross_modal_from_meter":false,
|
33 |
+
"patch_size":16,
|
34 |
+
"resolution_before":224,
|
35 |
+
"stop_gradient":false,
|
36 |
+
"size":288,
|
37 |
+
"task_head_layers":2,
|
38 |
+
"test_only":false,
|
39 |
+
"tokenizer":"roberta-base",
|
40 |
+
"unfreeze_RoBERTa_attention":false,
|
41 |
+
"unfreeze_RoBERTa_embeddings":false,
|
42 |
+
"unfreeze_RoBERTa_encoder":false,
|
43 |
+
"unfreeze_RoBERTa_layernorm":false,
|
44 |
+
"unfreeze_ViT_attention":false,
|
45 |
+
"unfreeze_ViT_layernorm":false,
|
46 |
+
"vit":"ViT-B/16",
|
47 |
+
"vit_layernorm_init_from_vit":false,
|
48 |
+
"vit_layernorm_shared":true,
|
49 |
+
"vit_remove_last":false,
|
50 |
+
"vocab_size":50265
|
|
|
|
|
51 |
}
|