fozziethebeat commited on
Commit
c97698e
1 Parent(s): 10e5f23

Upload config

Browse files
Files changed (1) hide show
  1. config.json +22 -4
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_name_or_path": "./checkpoints/vicuna-7b-v1-5",
3
  "architectures": [
4
- "LlavaMistralForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
@@ -11,6 +11,7 @@
11
  "freeze_mm_vision_resampler": false,
12
  "hidden_act": "silu",
13
  "hidden_size": 4096,
 
14
  "image_aspect_ratio": "anyres",
15
  "image_crop_resolution": 224,
16
  "image_grid_pinpoints": [
@@ -36,6 +37,7 @@
36
  ]
37
  ],
38
  "image_split_resolution": 224,
 
39
  "initializer_range": 0.02,
40
  "intermediate_size": 11008,
41
  "max_position_embeddings": 4096,
@@ -50,16 +52,19 @@
50
  "mm_vision_select_layer": -2,
51
  "mm_vision_tower": "openai/clip-vit-large-patch14-336",
52
  "mm_vision_tower_lr": 2e-06,
53
- "model_type": "llava_mistral",
54
  "num_attention_heads": 32,
55
  "num_hidden_layers": 32,
56
  "num_key_value_heads": 32,
57
- "pad_token_id": 0,
58
  "pretraining_tp": 1,
 
59
  "rms_norm_eps": 1e-05,
60
  "rope_scaling": null,
61
  "rope_theta": 10000.0,
62
- "sliding_window": 4096,
 
 
63
  "tie_word_embeddings": false,
64
  "tokenizer_model_max_length": 4096,
65
  "tokenizer_padding_side": "right",
@@ -70,5 +75,18 @@
70
  "unfreeze_mm_vision_tower": true,
71
  "use_cache": true,
72
  "use_mm_proj": true,
 
 
 
 
 
 
 
 
 
 
 
 
 
73
  "vocab_size": 32000
74
  }
 
1
  {
2
  "_name_or_path": "./checkpoints/vicuna-7b-v1-5",
3
  "architectures": [
4
+ "LlavaLlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
 
11
  "freeze_mm_vision_resampler": false,
12
  "hidden_act": "silu",
13
  "hidden_size": 4096,
14
+ "ignore_index": -100,
15
  "image_aspect_ratio": "anyres",
16
  "image_crop_resolution": 224,
17
  "image_grid_pinpoints": [
 
37
  ]
38
  ],
39
  "image_split_resolution": 224,
40
+ "image_token_index": 32000,
41
  "initializer_range": 0.02,
42
  "intermediate_size": 11008,
43
  "max_position_embeddings": 4096,
 
52
  "mm_vision_select_layer": -2,
53
  "mm_vision_tower": "openai/clip-vit-large-patch14-336",
54
  "mm_vision_tower_lr": 2e-06,
55
+ "model_type": "llava",
56
  "num_attention_heads": 32,
57
  "num_hidden_layers": 32,
58
  "num_key_value_heads": 32,
59
+ "pad_token_id": 32001,
60
  "pretraining_tp": 1,
61
+ "projector_hidden_act": "gelu",
62
  "rms_norm_eps": 1e-05,
63
  "rope_scaling": null,
64
  "rope_theta": 10000.0,
65
+ "text_config": {
66
+ "model_type": "llama"
67
+ },
68
  "tie_word_embeddings": false,
69
  "tokenizer_model_max_length": 4096,
70
  "tokenizer_padding_side": "right",
 
75
  "unfreeze_mm_vision_tower": true,
76
  "use_cache": true,
77
  "use_mm_proj": true,
78
+ "vision_config": {
79
+ "hidden_size": 1024,
80
+ "image_size": 336,
81
+ "intermediate_size": 4096,
82
+ "model_type": "clip_vision_model",
83
+ "num_attention_heads": 16,
84
+ "num_hidden_layers": 24,
85
+ "patch_size": 14,
86
+ "projection_dim": 768,
87
+ "vocab_size": 32000
88
+ },
89
+ "vision_feature_layer": -2,
90
+ "vision_feature_select_strategy": "default",
91
  "vocab_size": 32000
92
  }