codeShare commited on
Commit
e1a4ba4
·
verified ·
1 Parent(s): aeb9967

Delete config.json

Browse files
Files changed (1) hide show
  1. config.json +0 -88
config.json DELETED
@@ -1,88 +0,0 @@
1
- {
2
- "architectures": [
3
- "LlavaForConditionalGeneration"
4
- ],
5
- "dtype": "bfloat16",
6
- "image_seq_length": 729,
7
- "image_token_index": 128077,
8
- "model_type": "llava",
9
- "multimodal_projector_bias": true,
10
- "projector_hidden_act": "gelu",
11
- "quantization_config": {
12
- "_load_in_4bit": true,
13
- "_load_in_8bit": false,
14
- "bnb_4bit_compute_dtype": "float16",
15
- "bnb_4bit_quant_storage": "uint8",
16
- "bnb_4bit_quant_type": "nf4",
17
- "bnb_4bit_use_double_quant": false,
18
- "llm_int8_enable_fp32_cpu_offload": false,
19
- "llm_int8_has_fp16_weight": false,
20
- "llm_int8_skip_modules": null,
21
- "llm_int8_threshold": 6.0,
22
- "load_in_4bit": true,
23
- "load_in_8bit": false,
24
- "quant_method": "bitsandbytes"
25
- },
26
- "text_config": {
27
- "_name_or_path": "meta-llama/Llama-3.1-8B-Instruct",
28
- "architectures": [
29
- "LlamaForCausalLM"
30
- ],
31
- "attention_bias": false,
32
- "attention_dropout": 0.0,
33
- "bos_token_id": 128000,
34
- "dtype": "bfloat16",
35
- "eos_token_id": [
36
- 128001,
37
- 128008,
38
- 128009
39
- ],
40
- "head_dim": 128,
41
- "hidden_act": "silu",
42
- "hidden_size": 4096,
43
- "initializer_range": 0.02,
44
- "intermediate_size": 14336,
45
- "max_position_embeddings": 131072,
46
- "mlp_bias": false,
47
- "model_type": "llama",
48
- "num_attention_heads": 32,
49
- "num_hidden_layers": 32,
50
- "num_key_value_heads": 8,
51
- "pad_token_id": null,
52
- "pretraining_tp": 1,
53
- "rms_norm_eps": 1e-05,
54
- "rope_parameters": {
55
- "factor": 8.0,
56
- "high_freq_factor": 4.0,
57
- "low_freq_factor": 1.0,
58
- "original_max_position_embeddings": 8192,
59
- "rope_theta": 500000.0,
60
- "rope_type": "llama3"
61
- },
62
- "tie_word_embeddings": false,
63
- "use_cache": true,
64
- "vocab_size": 128256
65
- },
66
- "tie_word_embeddings": false,
67
- "transformers_version": "5.0.0",
68
- "vision_config": {
69
- "_name_or_path": "google/siglip2-so400m-patch14-384",
70
- "architectures": [
71
- "SiglipVisionModel"
72
- ],
73
- "attention_dropout": 0.0,
74
- "dtype": "bfloat16",
75
- "hidden_act": "gelu_pytorch_tanh",
76
- "hidden_size": 1152,
77
- "image_size": 384,
78
- "intermediate_size": 4304,
79
- "layer_norm_eps": 1e-06,
80
- "model_type": "siglip_vision_model",
81
- "num_attention_heads": 16,
82
- "num_channels": 3,
83
- "num_hidden_layers": 27,
84
- "patch_size": 14
85
- },
86
- "vision_feature_layer": -2,
87
- "vision_feature_select_strategy": "full"
88
- }