prince-canuma commited on
Commit
8414b69
1 Parent(s): 5477028

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: deepseek
4
+ license_link: https://github.com/deepseek-ai/DeepSeek-LLM/blob/HEAD/LICENSE-MODEL
5
+ pipeline_tag: image-text-to-text
6
+ library_name: transformers
7
+ tags:
8
+ - mlx
9
+ ---
10
+
11
+ # mlx-community/deepseek-vl2-small-6bit
12
+ This model was converted to MLX format from [`prince-canuma/deepseek-vl2-small`]() using mlx-vlm version **0.1.5**.
13
+ Refer to the [original model card](https://huggingface.co/prince-canuma/deepseek-vl2-small) for more details on the model.
14
+ ## Use with mlx
15
+
16
+ ```bash
17
+ pip install -U mlx-vlm
18
+ ```
19
+
20
+ ```bash
21
+ python -m mlx_vlm.generate --model mlx-community/deepseek-vl2-small-6bit --max-tokens 100 --temp 0.0
22
+ ```
chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}<|User|>:{% elif message['role'] == 'assistant' %}<|Assistant|>{% endif %} {{message['content']}}\n\n{% endfor %}{% if add_generation_prompt %}<|Assistant|>:{% endif %}"
3
+ }
config.json ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "candidate_resolutions": [
3
+ [
4
+ 384,
5
+ 384
6
+ ],
7
+ [
8
+ 384,
9
+ 768
10
+ ],
11
+ [
12
+ 768,
13
+ 384
14
+ ],
15
+ [
16
+ 384,
17
+ 1152
18
+ ],
19
+ [
20
+ 1152,
21
+ 384
22
+ ],
23
+ [
24
+ 384,
25
+ 1536
26
+ ],
27
+ [
28
+ 1536,
29
+ 384
30
+ ],
31
+ [
32
+ 768,
33
+ 768
34
+ ],
35
+ [
36
+ 384,
37
+ 1920
38
+ ],
39
+ [
40
+ 1920,
41
+ 384
42
+ ],
43
+ [
44
+ 384,
45
+ 2304
46
+ ],
47
+ [
48
+ 2304,
49
+ 384
50
+ ],
51
+ [
52
+ 768,
53
+ 1152
54
+ ],
55
+ [
56
+ 1152,
57
+ 768
58
+ ],
59
+ [
60
+ 384,
61
+ 2688
62
+ ],
63
+ [
64
+ 2688,
65
+ 384
66
+ ],
67
+ [
68
+ 384,
69
+ 3072
70
+ ],
71
+ [
72
+ 3072,
73
+ 384
74
+ ],
75
+ [
76
+ 768,
77
+ 1536
78
+ ],
79
+ [
80
+ 1536,
81
+ 768
82
+ ],
83
+ [
84
+ 384,
85
+ 3456
86
+ ],
87
+ [
88
+ 3456,
89
+ 384
90
+ ],
91
+ [
92
+ 1152,
93
+ 1152
94
+ ]
95
+ ],
96
+ "global_view_pos": "head",
97
+ "language_config": {
98
+ "architectures": [
99
+ "DeepseekV2ForCausalLM"
100
+ ],
101
+ "auto_map": {
102
+ "AutoConfig": "configuration_deepseek.DeepseekV2Config",
103
+ "AutoModel": "modeling_deepseek.DeepseekV2Model",
104
+ "AutoModelForCausalLM": "modeling_deepseek.DeepseekV2ForCausalLM"
105
+ },
106
+ "first_k_dense_replace": 1,
107
+ "hidden_size": 2048,
108
+ "intermediate_size": 10944,
109
+ "lm_head": true,
110
+ "max_position_embeddings": 4096,
111
+ "model_type": "deepseek_v2",
112
+ "moe_intermediate_size": 1408,
113
+ "n_group": 1,
114
+ "n_routed_experts": 64,
115
+ "n_shared_experts": 2,
116
+ "num_attention_heads": 16,
117
+ "num_experts_per_tok": 6,
118
+ "num_hidden_layers": 27,
119
+ "num_key_value_heads": 16,
120
+ "q_lora_rank": null,
121
+ "rm_head": false,
122
+ "topk_group": 1,
123
+ "topk_method": "greedy",
124
+ "torch_dtype": "bfloat16"
125
+ },
126
+ "model_type": "deepseek_vl_v2",
127
+ "projector_config": {
128
+ "model_type": "mlp_projector"
129
+ },
130
+ "quantization": {
131
+ "group_size": 64,
132
+ "bits": 6
133
+ },
134
+ "tile_tag": "2D",
135
+ "torch_dtype": "bfloat16",
136
+ "transformers_version": "4.38.2",
137
+ "vision_config": {
138
+ "layers": 27,
139
+ "mlp_ratio": 3.7362,
140
+ "model_name": "siglip_so400m_patch14_384",
141
+ "model_type": "vision",
142
+ "patch_size": 14,
143
+ "width": 1152,
144
+ "skip_vision_non_divisible": true
145
+ }
146
+ }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a759f955eeb9f3de44d4f8301bb5ae3ed35753e437508144cafb0514aa13228b
3
+ size 5340637305
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd0572961dd240850b087701a2302c526f0d418fe8971c950e508965fc14a7f5
3
+ size 5367313527
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2e9b137b4a2292ac59592219ac8ff25439e4922d3a52715585f4e9a1de4e5ef
3
+ size 2584128337
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
processor_config.json ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_special_token": false,
3
+ "candidate_resolutions": [
4
+ [
5
+ 384,
6
+ 384
7
+ ],
8
+ [
9
+ 384,
10
+ 768
11
+ ],
12
+ [
13
+ 768,
14
+ 384
15
+ ],
16
+ [
17
+ 384,
18
+ 1152
19
+ ],
20
+ [
21
+ 1152,
22
+ 384
23
+ ],
24
+ [
25
+ 384,
26
+ 1536
27
+ ],
28
+ [
29
+ 1536,
30
+ 384
31
+ ],
32
+ [
33
+ 768,
34
+ 768
35
+ ],
36
+ [
37
+ 384,
38
+ 1920
39
+ ],
40
+ [
41
+ 1920,
42
+ 384
43
+ ],
44
+ [
45
+ 384,
46
+ 2304
47
+ ],
48
+ [
49
+ 2304,
50
+ 384
51
+ ],
52
+ [
53
+ 768,
54
+ 1152
55
+ ],
56
+ [
57
+ 1152,
58
+ 768
59
+ ],
60
+ [
61
+ 384,
62
+ 2688
63
+ ],
64
+ [
65
+ 2688,
66
+ 384
67
+ ],
68
+ [
69
+ 384,
70
+ 3072
71
+ ],
72
+ [
73
+ 3072,
74
+ 384
75
+ ],
76
+ [
77
+ 768,
78
+ 1536
79
+ ],
80
+ [
81
+ 1536,
82
+ 768
83
+ ],
84
+ [
85
+ 384,
86
+ 3456
87
+ ],
88
+ [
89
+ 3456,
90
+ 384
91
+ ],
92
+ [
93
+ 1152,
94
+ 1152
95
+ ]
96
+ ],
97
+ "downsample_ratio": 2,
98
+ "ignore_id": -100,
99
+ "image_mean": [
100
+ 0.5,
101
+ 0.5,
102
+ 0.5
103
+ ],
104
+ "image_std": [
105
+ 0.5,
106
+ 0.5,
107
+ 0.5
108
+ ],
109
+ "image_token": "<image>",
110
+ "mask_prompt": false,
111
+ "normalize": true,
112
+ "pad_token": "<\uff5c\u2581pad\u2581\uff5c>",
113
+ "patch_size": 14,
114
+ "processor_class": "DeepseekVLV2Processor",
115
+ "sft_format": "deepseek"
116
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|User|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|Assistant|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ }
17
+ ],
18
+ "bos_token": {
19
+ "content": "<|begin▁of▁sentence|>",
20
+ "lstrip": false,
21
+ "normalized": true,
22
+ "rstrip": false,
23
+ "single_word": false
24
+ },
25
+ "eos_token": {
26
+ "content": "<|end▁of▁sentence|>",
27
+ "lstrip": false,
28
+ "normalized": true,
29
+ "rstrip": false,
30
+ "single_word": false
31
+ },
32
+ "pad_token": {
33
+ "content": "<|▁pad▁|>",
34
+ "lstrip": false,
35
+ "normalized": false,
36
+ "rstrip": false,
37
+ "single_word": false
38
+ }
39
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "100000": {
7
+ "content": "<|begin▁of▁sentence|>",
8
+ "lstrip": false,
9
+ "normalized": true,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "100001": {
15
+ "content": "<|end▁of▁sentence|>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "100002": {
23
+ "content": "<|▁pad▁|>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "100003": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "100004": {
39
+ "content": "<|ref|>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ },
46
+ "100005": {
47
+ "content": "<|/ref|>",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false,
52
+ "special": true
53
+ },
54
+ "100006": {
55
+ "content": "<|det|>",
56
+ "lstrip": false,
57
+ "normalized": false,
58
+ "rstrip": false,
59
+ "single_word": false,
60
+ "special": true
61
+ },
62
+ "100007": {
63
+ "content": "<|/det|>",
64
+ "lstrip": false,
65
+ "normalized": false,
66
+ "rstrip": false,
67
+ "single_word": false,
68
+ "special": true
69
+ },
70
+ "100008": {
71
+ "content": "<|grounding|>",
72
+ "lstrip": false,
73
+ "normalized": false,
74
+ "rstrip": false,
75
+ "single_word": false,
76
+ "special": true
77
+ },
78
+ "100009": {
79
+ "content": "<|User|>",
80
+ "lstrip": false,
81
+ "normalized": false,
82
+ "rstrip": false,
83
+ "single_word": false,
84
+ "special": true
85
+ },
86
+ "100010": {
87
+ "content": "<|Assistant|>",
88
+ "lstrip": false,
89
+ "normalized": false,
90
+ "rstrip": false,
91
+ "single_word": false,
92
+ "special": true
93
+ }
94
+ },
95
+ "additional_special_tokens": [
96
+ "<|User|>",
97
+ "<|Assistant|>"
98
+ ],
99
+ "bos_token": "<|begin▁of▁sentence|>",
100
+ "chat_template": "{%- set found_item = false -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set found_item = true -%}\n {%- endif -%}\n{%- endfor -%}\n{%- if not found_item -%}\n{{'You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer.\\n'}}\n{%- endif %}\n{%- for message in messages %}\n {%- if message['role'] == 'system' %}\n{{ message['content'] }}\n {%- else %}\n {%- if message['role'] == 'user' %}\n{{'### Instruction:\\n' + message['content'] + '\\n'}}\n {%- else %}\n{{'### Response:\\n' + message['content'] + '\\n<|EOT|>\\n'}}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{{'### Response:\\n'}}\n",
101
+ "clean_up_tokenization_spaces": false,
102
+ "eos_token": "<|end▁of▁sentence|>",
103
+ "extra_special_tokens": {},
104
+ "legacy": true,
105
+ "model_max_length": 100000000,
106
+ "pad_token": "<|▁pad▁|>",
107
+ "processor_class": "DeepseekVLV2Processor",
108
+ "sp_model_kwargs": {},
109
+ "tokenizer_class": "LlamaTokenizer",
110
+ "unk_token": null,
111
+ "use_default_system_prompt": false
112
+ }