stanrom commited on
Commit
a41294f
1 Parent(s): 296c480

Upload config.json

Browse files
Files changed (1) hide show
  1. config.json +21 -27
config.json CHANGED
@@ -1,42 +1,36 @@
1
  {
2
- "_name_or_path": "ShareGPT4V-7B",
3
  "architectures": [
4
- "Share4VLlamaForCausalLM"
5
  ],
 
 
 
 
 
 
6
  "bos_token_id": 1,
 
7
  "eos_token_id": 2,
8
- "freeze_mm_mlp_adapter": false,
9
  "hidden_act": "silu",
10
  "hidden_size": 4096,
11
- "image_aspect_ratio": "pad",
12
- "image_grid_pinpoints": null,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 11008,
15
- "max_position_embeddings": 4096,
16
- "mm_hidden_size": 1024,
17
- "mm_projector_lr": null,
18
- "mm_projector_type": "mlp2x_gelu",
19
- "mm_use_im_patch_token": false,
20
- "mm_use_im_start_end": false,
21
- "mm_vision_select_feature": "patch",
22
- "mm_vision_select_layer": -2,
23
- "mm_vision_tower": "Lin-Chen/ShareGPT4V-7B_Pretrained_vit-large336-l12",
24
- "model_type": "share4v",
25
  "num_attention_heads": 32,
26
  "num_hidden_layers": 32,
27
- "num_key_value_heads": 32,
28
- "pad_token_id": 0,
29
- "pretraining_tp": 1,
30
  "rms_norm_eps": 1e-05,
31
- "rope_scaling": null,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
- "transformers_version": "4.31.0",
35
- "tune_entire_model": false,
36
- "tune_mm_mlp_adapter": false,
37
- "tune_vision_tower": false,
38
- "use_cache": true,
39
- "use_mm_proj": true,
40
- "vision_tower_lr": null,
41
- "vocab_size": 32000
42
  }
 
1
  {
2
+ "_name_or_path": "/mnt/petrelfs/dongxiaoyi/gittest/IXC/output/DS_full_044HD1_1128_BS256/checkpoint-714",
3
  "architectures": [
4
+ "InternLMXComposerForCausalLM"
5
  ],
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_InternLM_XComposer.InternLMXComposerConfig",
8
+ "AutoModel": "modeling_InternLM_XComposer.InternLMXComposerForCausalLM",
9
+ "AutoModelForCausalLM": "modeling_InternLM_XComposer.InternLMXComposerForCausalLM"
10
+ },
11
+ "bias": true,
12
  "bos_token_id": 1,
13
+ "device": "cuda",
14
  "eos_token_id": 2,
 
15
  "hidden_act": "silu",
16
  "hidden_size": 4096,
 
 
17
  "initializer_range": 0.02,
18
  "intermediate_size": 11008,
19
+ "intern_converted_llm": true,
20
+ "internlm_lora": null,
21
+ "kqvo_bias": true,
22
+ "max_length": 768,
23
+ "max_position_embeddings": 2048,
24
+ "model_type": "InternLMXComposer",
 
 
 
 
25
  "num_attention_heads": 32,
26
  "num_hidden_layers": 32,
27
+ "num_quant": 32,
28
+ "num_query_token": 64,
29
+ "pad_token_id": -1,
30
  "rms_norm_eps": 1e-05,
 
31
  "tie_word_embeddings": false,
32
  "torch_dtype": "bfloat16",
33
+ "transformers_version": "4.33.1",
34
+ "use_cache": false,
35
+ "vocab_size": 103172
 
 
 
 
 
36
  }