File size: 3,158 Bytes
983f690
fc837bd
 
 
 
 
8772eec
983f690
 
 
fc837bd
983f690
fc837bd
983f690
 
 
 
fc837bd
983f690
fc837bd
 
983f690
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fc837bd
 
983f690
 
 
 
 
fc837bd
983f690
 
 
 
 
 
fc837bd
983f690
 
fc837bd
983f690
fc837bd
983f690
fc837bd
983f690
 
fc837bd
 
983f690
 
 
 
 
fc837bd
983f690
 
 
fc837bd
983f690
 
 
 
 
fc837bd
 
983f690
 
 
 
 
 
 
 
 
 
fc837bd
983f690
 
 
fc837bd
983f690
 
fc837bd
 
 
 
 
 
 
 
 
 
983f690
 
fc837bd
983f690
fc837bd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
{
  "auto_map": {
    "AutoConfig": "modeling_llava.LlavaConfig",
    "AutoModel": "modeling_llava.LlavaForCausalLM",
    "AutoModelForCausalLM": "modeling_llava.LlavaForCausalLM"
  },
  "model_type": "HelpingAI",
  "ignore_index": -100,
  "image_token_index": 50297,
  "projector_hidden_act": "gelu",
  "projector_tokens_num": 1,
  "text_config": {
    "_name_or_path": "vince62s/phi-2-psy",
    "add_cross_attention": false,
    "architectures": [
      "PhiForCausalLM"
    ],
    "attention_dropout": 0.0,
    "auto_map": {
      "AutoConfig": "vince62s/phi-2-psy--configuration_phi.PhiConfig",
      "AutoModelForCausalLM": "vince62s/phi-2-psy--modeling_phi.PhiForCausalLM"
    },
    "bad_words_ids": null,
    "begin_suppress_tokens": null,
    "bos_token_id": null,
    "chunk_size_feed_forward": 0,
    "cross_attention_hidden_size": null,
    "decoder_start_token_id": null,
    "diversity_penalty": 0.0,
    "do_sample": false,
    "early_stopping": false,
    "embd_pdrop": 0.0,
    "encoder_no_repeat_ngram_size": 0,
    "eos_token_id": null,
    "exponential_decay_length_penalty": null,
    "finetuning_task": null,
    "forced_bos_token_id": null,
    "forced_eos_token_id": null,
    "hidden_act": "gelu_new",
    "hidden_size": 2560,
    "id2label": {
      "0": "LABEL_0",
      "1": "LABEL_1"
    },
    "initializer_range": 0.02,
    "intermediate_size": 10240,
    "is_decoder": false,
    "is_encoder_decoder": false,
    "label2id": {
      "LABEL_0": 0,
      "LABEL_1": 1
    },
    "layer_norm_eps": 1e-05,
    "length_penalty": 1.0,
    "max_length": 20,
    "max_position_embeddings": 2048,
    "min_length": 0,
    "model_type": "phi",
    "no_repeat_ngram_size": 0,
    "num_attention_heads": 32,
    "num_beam_groups": 1,
    "num_beams": 1,
    "num_hidden_layers": 32,
    "num_key_value_heads": 32,
    "num_return_sequences": 1,
    "output_attentions": false,
    "output_hidden_states": false,
    "output_scores": false,
    "pad_token_id": null,
    "partial_rotary_factor": 0.4,
    "prefix": null,
    "problem_type": null,
    "pruned_heads": {},
    "qk_layernorm": false,
    "remove_invalid_values": false,
    "repetition_penalty": 1.0,
    "resid_pdrop": 0.1,
    "return_dict": true,
    "return_dict_in_generate": false,
    "rope_scaling": null,
    "rope_theta": 10000.0,
    "sep_token_id": null,
    "suppress_tokens": null,
    "task_specific_params": null,
    "temperature": 1.0,
    "tf_legacy_loss": false,
    "tie_encoder_decoder": false,
    "tie_word_embeddings": false,
    "tokenizer_class": null,
    "top_k": 50,
    "top_p": 1.0,
    "torch_dtype": "bfloat16",
    "torchscript": false,
    "typical_p": 1.0,
    "use_bfloat16": false,
    "use_cache": true,
    "vocab_size": 51200
  },
  "torch_dtype": "bfloat16",
  "transformers_version": "4.37.2",
  "vision_config": {
    "hidden_size": 1152,
    "image_size": 384,
    "intermediate_size": 4304,
    "model_type": "siglip_vision_model",
    "num_attention_heads": 16,
    "num_hidden_layers": 27,
    "patch_size": 14
  },
  "vision_embed_dim": 1152,
  "vision_tower_name": "google/siglip-so400m-patch14-384",
  "vocab_size": 51200
}