Xenova HF staff commited on
Commit
438e7c6
·
verified ·
1 Parent(s): 3d24691

Create config.json

Browse files
Files changed (1) hide show
  1. config.json +64 -0
config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aligner_config": {
3
+ "cls": "MlpProjector",
4
+ "model_type": "aligner",
5
+ "params": {
6
+ "depth": 2,
7
+ "input_dim": 1024,
8
+ "n_embed": 2048,
9
+ "projector_type": "mlp_gelu"
10
+ }
11
+ },
12
+ "gen_aligner_config": {
13
+ "cls": "MlpProjector",
14
+ "model_type": "gen_aligner",
15
+ "params": {
16
+ "depth": 2,
17
+ "input_dim": 8,
18
+ "n_embed": 2048,
19
+ "projector_type": "mlp_gelu"
20
+ }
21
+ },
22
+ "gen_head_config": {
23
+ "cls": "vision_head",
24
+ "model_type": "gen_head",
25
+ "params": {
26
+ "image_token_embed": 2048,
27
+ "image_token_size": 16384,
28
+ "n_embed": 2048
29
+ }
30
+ },
31
+ "gen_vision_config": {
32
+ "cls": "VQ-16",
33
+ "model_type": "gen_vision",
34
+ "params": {
35
+ "image_token_size": 16384,
36
+ "n_embed": 8
37
+ }
38
+ },
39
+ "language_config": {
40
+ "hidden_size": 2048,
41
+ "intermediate_size": 5632,
42
+ "max_position_embeddings": 16384,
43
+ "model_type": "llama",
44
+ "num_attention_heads": 16,
45
+ "num_hidden_layers": 24,
46
+ "num_key_value_heads": 16,
47
+ "torch_dtype": "bfloat16",
48
+ "vocab_size": 102400,
49
+ "_attn_implementation": "flash_attention_2"
50
+ },
51
+ "model_type": "multi_modality",
52
+ "torch_dtype": "bfloat16",
53
+ "transformers_version": "4.38.2",
54
+ "vision_config": {
55
+ "cls": "CLIPVisionTower",
56
+ "model_type": "vision",
57
+ "params": {
58
+ "image_size": 384,
59
+ "model_name": "siglip_large_patch16_384",
60
+ "select_feature": "same",
61
+ "select_layer": -1
62
+ }
63
+ }
64
+ }