prince-canuma
commited on
Commit
•
82b38f2
1
Parent(s):
1850355
Upload 10 files
Browse files- README.md +2 -5
- model.safetensors +2 -2
- model.safetensors.index.json +5 -1
- tokenizer.json +1 -0
- tokenizer_config.json +0 -1
README.md
CHANGED
@@ -11,7 +11,7 @@ arxiv: 2304.08485
|
|
11 |
---
|
12 |
|
13 |
# mlx-community/llava-1.5-7b-4bit
|
14 |
-
This model was converted to MLX format from [`llava-hf/llava-1.5-7b-hf`]() using mlx-
|
15 |
Refer to the [original model card](https://huggingface.co/llava-hf/llava-1.5-7b-hf) for more details on the model.
|
16 |
## Use with mlx
|
17 |
|
@@ -20,8 +20,5 @@ pip install -U mlx-vlm
|
|
20 |
```
|
21 |
|
22 |
```bash
|
23 |
-
python -m mlx_vlm.generate --model mlx-community/llava-1.5-7b-4bit
|
24 |
-
--prompt "what are these?" --image "http://images.cocodataset.org/val2017/000000039769.jpg" \
|
25 |
-
--max-tokens 100 --temp 0.0
|
26 |
```
|
27 |
-
|
|
|
11 |
---
|
12 |
|
13 |
# mlx-community/llava-1.5-7b-4bit
|
14 |
+
This model was converted to MLX format from [`llava-hf/llava-1.5-7b-hf`]() using mlx-vlm version **0.0.4**.
|
15 |
Refer to the [original model card](https://huggingface.co/llava-hf/llava-1.5-7b-hf) for more details on the model.
|
16 |
## Use with mlx
|
17 |
|
|
|
20 |
```
|
21 |
|
22 |
```bash
|
23 |
+
python -m mlx_vlm.generate --model mlx-community/llava-1.5-7b-4bit --max-tokens 100 --temp 0.0
|
|
|
|
|
24 |
```
|
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13c021a1fad60c7fcd9e3f3ff1c7150fbe6a595151b5d5e979dbdb0122586ae8
|
3 |
+
size 3975094479
|
model.safetensors.index.json
CHANGED
@@ -1,11 +1,13 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"language_model.lm_head.biases": "model.safetensors",
|
7 |
"language_model.lm_head.scales": "model.safetensors",
|
8 |
"language_model.lm_head.weight": "model.safetensors",
|
|
|
|
|
9 |
"language_model.model.embed_tokens.weight": "model.safetensors",
|
10 |
"language_model.model.layers.0.input_layernorm.weight": "model.safetensors",
|
11 |
"language_model.model.layers.0.mlp.down_proj.biases": "model.safetensors",
|
@@ -754,6 +756,8 @@
|
|
754 |
"multi_modal_projector.linear_2.weight": "model.safetensors",
|
755 |
"vision_tower.vision_model.embeddings.class_embedding": "model.safetensors",
|
756 |
"vision_tower.vision_model.embeddings.patch_embedding.weight": "model.safetensors",
|
|
|
|
|
757 |
"vision_tower.vision_model.embeddings.position_embedding.weight": "model.safetensors",
|
758 |
"vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "model.safetensors",
|
759 |
"vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "model.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 3974904384
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"language_model.lm_head.biases": "model.safetensors",
|
7 |
"language_model.lm_head.scales": "model.safetensors",
|
8 |
"language_model.lm_head.weight": "model.safetensors",
|
9 |
+
"language_model.model.embed_tokens.biases": "model.safetensors",
|
10 |
+
"language_model.model.embed_tokens.scales": "model.safetensors",
|
11 |
"language_model.model.embed_tokens.weight": "model.safetensors",
|
12 |
"language_model.model.layers.0.input_layernorm.weight": "model.safetensors",
|
13 |
"language_model.model.layers.0.mlp.down_proj.biases": "model.safetensors",
|
|
|
756 |
"multi_modal_projector.linear_2.weight": "model.safetensors",
|
757 |
"vision_tower.vision_model.embeddings.class_embedding": "model.safetensors",
|
758 |
"vision_tower.vision_model.embeddings.patch_embedding.weight": "model.safetensors",
|
759 |
+
"vision_tower.vision_model.embeddings.position_embedding.biases": "model.safetensors",
|
760 |
+
"vision_tower.vision_model.embeddings.position_embedding.scales": "model.safetensors",
|
761 |
"vision_tower.vision_model.embeddings.position_embedding.weight": "model.safetensors",
|
762 |
"vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "model.safetensors",
|
763 |
"vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "model.safetensors",
|
tokenizer.json
CHANGED
@@ -152,6 +152,7 @@
|
|
152 |
"end_of_word_suffix": null,
|
153 |
"fuse_unk": true,
|
154 |
"byte_fallback": true,
|
|
|
155 |
"vocab": {
|
156 |
"<unk>": 0,
|
157 |
"<s>": 1,
|
|
|
152 |
"end_of_word_suffix": null,
|
153 |
"fuse_unk": true,
|
154 |
"byte_fallback": true,
|
155 |
+
"ignore_merges": false,
|
156 |
"vocab": {
|
157 |
"<unk>": 0,
|
158 |
"<s>": 1,
|
tokenizer_config.json
CHANGED
@@ -46,7 +46,6 @@
|
|
46 |
"bos_token": "<s>",
|
47 |
"clean_up_tokenization_spaces": false,
|
48 |
"eos_token": "</s>",
|
49 |
-
"legacy": false,
|
50 |
"model_max_length": 1000000000000000019884624838656,
|
51 |
"pad_token": "<pad>",
|
52 |
"padding_side": "left",
|
|
|
46 |
"bos_token": "<s>",
|
47 |
"clean_up_tokenization_spaces": false,
|
48 |
"eos_token": "</s>",
|
|
|
49 |
"model_max_length": 1000000000000000019884624838656,
|
50 |
"pad_token": "<pad>",
|
51 |
"padding_side": "left",
|