Abhaykoul commited on
Commit
fc837bd
1 Parent(s): 31a2295

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +38 -44
config.json CHANGED
@@ -1,24 +1,24 @@
1
  {
2
- "architectures": [
3
- "LlavaForConditionalGeneration"
4
- ],
 
 
 
5
  "ignore_index": -100,
6
  "image_token_index": 50297,
7
- "max_image_tokens": 100,
8
- "model_type": "llava",
9
  "projector_hidden_act": "gelu",
10
- "projector_tokens_num": 5,
11
  "text_config": {
12
- "_name_or_path": "cognitivecomputations/dolphin-2_6-phi-2",
13
- "activation_function": "gelu_new",
14
  "add_cross_attention": false,
15
  "architectures": [
16
  "PhiForCausalLM"
17
  ],
18
- "attn_pdrop": 0.0,
19
  "auto_map": {
20
- "AutoConfig": "cognitivecomputations/dolphin-2_6-phi-2--configuration_phi.PhiConfig",
21
- "AutoModelForCausalLM": "cognitivecomputations/dolphin-2_6-phi-2--modeling_phi.PhiForCausalLM"
22
  },
23
  "bad_words_ids": null,
24
  "begin_suppress_tokens": null,
@@ -34,51 +34,51 @@
34
  "eos_token_id": null,
35
  "exponential_decay_length_penalty": null,
36
  "finetuning_task": null,
37
- "flash_attn": false,
38
- "flash_rotary": false,
39
  "forced_bos_token_id": null,
40
  "forced_eos_token_id": null,
41
- "fused_dense": false,
 
42
  "id2label": {
43
  "0": "LABEL_0",
44
  "1": "LABEL_1"
45
  },
46
- "img_processor": null,
47
  "initializer_range": 0.02,
 
48
  "is_decoder": false,
49
  "is_encoder_decoder": false,
50
  "label2id": {
51
  "LABEL_0": 0,
52
  "LABEL_1": 1
53
  },
54
- "layer_norm_epsilon": 1e-05,
55
  "length_penalty": 1.0,
56
  "max_length": 20,
 
57
  "min_length": 0,
58
- "model_type": "phi-msft",
59
- "n_embd": 2560,
60
- "n_head": 32,
61
- "n_head_kv": null,
62
- "n_inner": null,
63
- "n_layer": 32,
64
- "n_positions": 2048,
65
  "no_repeat_ngram_size": 0,
 
66
  "num_beam_groups": 1,
67
  "num_beams": 1,
 
 
68
  "num_return_sequences": 1,
69
  "output_attentions": false,
70
  "output_hidden_states": false,
71
  "output_scores": false,
72
  "pad_token_id": null,
 
73
  "prefix": null,
74
  "problem_type": null,
75
  "pruned_heads": {},
 
76
  "remove_invalid_values": false,
77
  "repetition_penalty": 1.0,
78
  "resid_pdrop": 0.1,
79
  "return_dict": true,
80
  "return_dict_in_generate": false,
81
- "rotary_dim": 32,
 
82
  "sep_token_id": null,
83
  "suppress_tokens": null,
84
  "task_specific_params": null,
@@ -89,31 +89,25 @@
89
  "tokenizer_class": null,
90
  "top_k": 50,
91
  "top_p": 1.0,
92
- "torch_dtype": "float16",
93
  "torchscript": false,
94
  "typical_p": 1.0,
95
  "use_bfloat16": false,
96
- "use_cache": false,
97
  "vocab_size": 51200
98
  },
99
- "preprocess_config": {
100
- "mean": [
101
- 0.5,
102
- 0.5,
103
- 0.5
104
- ],
105
- "std": [
106
- 0.5,
107
- 0.5,
108
- 0.5
109
- ],
110
- "interpolation": "bicubic",
111
- "resize_mode": "squash",
112
- "size": 384
113
  },
114
- "torch_dtype": "float16",
115
- "transformers_version": "4.36.2",
116
  "vision_embed_dim": 1152,
117
- "vision_tower_name": "ViT-SO400M-14-SigLIP-384",
118
  "vocab_size": 51200
119
- }
 
1
  {
2
+ "auto_map": {
3
+ "AutoConfig": "modeling_llava.LlavaConfig",
4
+ "AutoModel": "modeling_llava.LlavaForCausalLM",
5
+ "AutoModelForCausalLM": "modeling_llava.LlavaForCausalLM"
6
+ },
7
+ "model_type": "mc-llava",
8
  "ignore_index": -100,
9
  "image_token_index": 50297,
 
 
10
  "projector_hidden_act": "gelu",
11
+ "projector_tokens_num": 1,
12
  "text_config": {
13
+ "_name_or_path": "vince62s/phi-2-psy",
 
14
  "add_cross_attention": false,
15
  "architectures": [
16
  "PhiForCausalLM"
17
  ],
18
+ "attention_dropout": 0.0,
19
  "auto_map": {
20
+ "AutoConfig": "vince62s/phi-2-psy--configuration_phi.PhiConfig",
21
+ "AutoModelForCausalLM": "vince62s/phi-2-psy--modeling_phi.PhiForCausalLM"
22
  },
23
  "bad_words_ids": null,
24
  "begin_suppress_tokens": null,
 
34
  "eos_token_id": null,
35
  "exponential_decay_length_penalty": null,
36
  "finetuning_task": null,
 
 
37
  "forced_bos_token_id": null,
38
  "forced_eos_token_id": null,
39
+ "hidden_act": "gelu_new",
40
+ "hidden_size": 2560,
41
  "id2label": {
42
  "0": "LABEL_0",
43
  "1": "LABEL_1"
44
  },
 
45
  "initializer_range": 0.02,
46
+ "intermediate_size": 10240,
47
  "is_decoder": false,
48
  "is_encoder_decoder": false,
49
  "label2id": {
50
  "LABEL_0": 0,
51
  "LABEL_1": 1
52
  },
53
+ "layer_norm_eps": 1e-05,
54
  "length_penalty": 1.0,
55
  "max_length": 20,
56
+ "max_position_embeddings": 2048,
57
  "min_length": 0,
58
+ "model_type": "phi",
 
 
 
 
 
 
59
  "no_repeat_ngram_size": 0,
60
+ "num_attention_heads": 32,
61
  "num_beam_groups": 1,
62
  "num_beams": 1,
63
+ "num_hidden_layers": 32,
64
+ "num_key_value_heads": 32,
65
  "num_return_sequences": 1,
66
  "output_attentions": false,
67
  "output_hidden_states": false,
68
  "output_scores": false,
69
  "pad_token_id": null,
70
+ "partial_rotary_factor": 0.4,
71
  "prefix": null,
72
  "problem_type": null,
73
  "pruned_heads": {},
74
+ "qk_layernorm": false,
75
  "remove_invalid_values": false,
76
  "repetition_penalty": 1.0,
77
  "resid_pdrop": 0.1,
78
  "return_dict": true,
79
  "return_dict_in_generate": false,
80
+ "rope_scaling": null,
81
+ "rope_theta": 10000.0,
82
  "sep_token_id": null,
83
  "suppress_tokens": null,
84
  "task_specific_params": null,
 
89
  "tokenizer_class": null,
90
  "top_k": 50,
91
  "top_p": 1.0,
92
+ "torch_dtype": "bfloat16",
93
  "torchscript": false,
94
  "typical_p": 1.0,
95
  "use_bfloat16": false,
96
+ "use_cache": true,
97
  "vocab_size": 51200
98
  },
99
+ "torch_dtype": "bfloat16",
100
+ "transformers_version": "4.37.2",
101
+ "vision_config": {
102
+ "hidden_size": 1152,
103
+ "image_size": 384,
104
+ "intermediate_size": 4304,
105
+ "model_type": "siglip_vision_model",
106
+ "num_attention_heads": 16,
107
+ "num_hidden_layers": 27,
108
+ "patch_size": 14
 
 
 
 
109
  },
 
 
110
  "vision_embed_dim": 1152,
111
+ "vision_tower_name": "google/siglip-so400m-patch14-384",
112
  "vocab_size": 51200
113
+ }