fozziethebeat commited on
Commit
f5d02da
1 Parent(s): 09cac3e

Upload config

Browse files
Files changed (1) hide show
  1. config.json +22 -3
config.json CHANGED
@@ -11,6 +11,7 @@
11
  "freeze_mm_vision_resampler": false,
12
  "hidden_act": "silu",
13
  "hidden_size": 4096,
 
14
  "image_aspect_ratio": "anyres",
15
  "image_crop_resolution": 224,
16
  "image_grid_pinpoints": [
@@ -36,6 +37,7 @@
36
  ]
37
  ],
38
  "image_split_resolution": 224,
 
39
  "initializer_range": 0.02,
40
  "intermediate_size": 11008,
41
  "max_position_embeddings": 4096,
@@ -50,24 +52,41 @@
50
  "mm_vision_select_layer": -2,
51
  "mm_vision_tower": "openai/clip-vit-large-patch14-336",
52
  "mm_vision_tower_lr": 2e-06,
53
- "model_type": "llava_llama",
54
  "num_attention_heads": 32,
55
  "num_hidden_layers": 32,
56
  "num_key_value_heads": 32,
57
- "pad_token_id": 0,
58
  "pretraining_tp": 1,
 
59
  "rms_norm_eps": 1e-05,
60
  "rope_scaling": null,
61
  "rope_theta": 10000.0,
 
 
 
62
  "tie_word_embeddings": false,
63
  "tokenizer_model_max_length": 4096,
64
  "tokenizer_padding_side": "right",
65
- "torch_dtype": "float16",
66
  "transformers_version": "4.37.2",
67
  "tune_mm_mlp_adapter": false,
68
  "tune_mm_vision_resampler": false,
69
  "unfreeze_mm_vision_tower": true,
70
  "use_cache": true,
71
  "use_mm_proj": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
72
  "vocab_size": 32000
73
  }
 
11
  "freeze_mm_vision_resampler": false,
12
  "hidden_act": "silu",
13
  "hidden_size": 4096,
14
+ "ignore_index": -100,
15
  "image_aspect_ratio": "anyres",
16
  "image_crop_resolution": 224,
17
  "image_grid_pinpoints": [
 
37
  ]
38
  ],
39
  "image_split_resolution": 224,
40
+ "image_token_index": 32000,
41
  "initializer_range": 0.02,
42
  "intermediate_size": 11008,
43
  "max_position_embeddings": 4096,
 
52
  "mm_vision_select_layer": -2,
53
  "mm_vision_tower": "openai/clip-vit-large-patch14-336",
54
  "mm_vision_tower_lr": 2e-06,
55
+ "model_type": "llava",
56
  "num_attention_heads": 32,
57
  "num_hidden_layers": 32,
58
  "num_key_value_heads": 32,
59
+ "pad_token_id": 32001,
60
  "pretraining_tp": 1,
61
+ "projector_hidden_act": "gelu",
62
  "rms_norm_eps": 1e-05,
63
  "rope_scaling": null,
64
  "rope_theta": 10000.0,
65
+ "text_config": {
66
+ "model_type": "llama"
67
+ },
68
  "tie_word_embeddings": false,
69
  "tokenizer_model_max_length": 4096,
70
  "tokenizer_padding_side": "right",
71
+ "torch_dtype": "bfloat16",
72
  "transformers_version": "4.37.2",
73
  "tune_mm_mlp_adapter": false,
74
  "tune_mm_vision_resampler": false,
75
  "unfreeze_mm_vision_tower": true,
76
  "use_cache": true,
77
  "use_mm_proj": true,
78
+ "vision_config": {
79
+ "hidden_size": 1024,
80
+ "image_size": 336,
81
+ "intermediate_size": 4096,
82
+ "model_type": "clip_vision_model",
83
+ "num_attention_heads": 16,
84
+ "num_hidden_layers": 24,
85
+ "patch_size": 14,
86
+ "projection_dim": 768,
87
+ "vocab_size": 32000
88
+ },
89
+ "vision_feature_layer": -2,
90
+ "vision_feature_select_strategy": "default",
91
  "vocab_size": 32000
92
  }