prince-canuma commited on
Commit
82b38f2
1 Parent(s): 1850355

Upload 10 files

Browse files
README.md CHANGED
@@ -11,7 +11,7 @@ arxiv: 2304.08485
11
  ---
12
 
13
  # mlx-community/llava-1.5-7b-4bit
14
- This model was converted to MLX format from [`llava-hf/llava-1.5-7b-hf`]() using mlx-vllm version **0.0.3**.
15
  Refer to the [original model card](https://huggingface.co/llava-hf/llava-1.5-7b-hf) for more details on the model.
16
  ## Use with mlx
17
 
@@ -20,8 +20,5 @@ pip install -U mlx-vlm
20
  ```
21
 
22
  ```bash
23
- python -m mlx_vlm.generate --model mlx-community/llava-1.5-7b-4bit \
24
- --prompt "what are these?" --image "http://images.cocodataset.org/val2017/000000039769.jpg" \
25
- --max-tokens 100 --temp 0.0
26
  ```
27
-
 
11
  ---
12
 
13
  # mlx-community/llava-1.5-7b-4bit
14
+ This model was converted to MLX format from [`llava-hf/llava-1.5-7b-hf`]() using mlx-vlm version **0.0.4**.
15
  Refer to the [original model card](https://huggingface.co/llava-hf/llava-1.5-7b-hf) for more details on the model.
16
  ## Use with mlx
17
 
 
20
  ```
21
 
22
  ```bash
23
+ python -m mlx_vlm.generate --model mlx-community/llava-1.5-7b-4bit --max-tokens 100 --temp 0.0
 
 
24
  ```
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0fbefdb9259abf4a2a438a057eaafa437d1b057f4b5b605b9211ba39ed8d237
3
- size 4164736149
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13c021a1fad60c7fcd9e3f3ff1c7150fbe6a595151b5d5e979dbdb0122586ae8
3
+ size 3975094479
model.safetensors.index.json CHANGED
@@ -1,11 +1,13 @@
1
  {
2
  "metadata": {
3
- "total_size": 4164546560
4
  },
5
  "weight_map": {
6
  "language_model.lm_head.biases": "model.safetensors",
7
  "language_model.lm_head.scales": "model.safetensors",
8
  "language_model.lm_head.weight": "model.safetensors",
 
 
9
  "language_model.model.embed_tokens.weight": "model.safetensors",
10
  "language_model.model.layers.0.input_layernorm.weight": "model.safetensors",
11
  "language_model.model.layers.0.mlp.down_proj.biases": "model.safetensors",
@@ -754,6 +756,8 @@
754
  "multi_modal_projector.linear_2.weight": "model.safetensors",
755
  "vision_tower.vision_model.embeddings.class_embedding": "model.safetensors",
756
  "vision_tower.vision_model.embeddings.patch_embedding.weight": "model.safetensors",
 
 
757
  "vision_tower.vision_model.embeddings.position_embedding.weight": "model.safetensors",
758
  "vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "model.safetensors",
759
  "vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "model.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 3974904384
4
  },
5
  "weight_map": {
6
  "language_model.lm_head.biases": "model.safetensors",
7
  "language_model.lm_head.scales": "model.safetensors",
8
  "language_model.lm_head.weight": "model.safetensors",
9
+ "language_model.model.embed_tokens.biases": "model.safetensors",
10
+ "language_model.model.embed_tokens.scales": "model.safetensors",
11
  "language_model.model.embed_tokens.weight": "model.safetensors",
12
  "language_model.model.layers.0.input_layernorm.weight": "model.safetensors",
13
  "language_model.model.layers.0.mlp.down_proj.biases": "model.safetensors",
 
756
  "multi_modal_projector.linear_2.weight": "model.safetensors",
757
  "vision_tower.vision_model.embeddings.class_embedding": "model.safetensors",
758
  "vision_tower.vision_model.embeddings.patch_embedding.weight": "model.safetensors",
759
+ "vision_tower.vision_model.embeddings.position_embedding.biases": "model.safetensors",
760
+ "vision_tower.vision_model.embeddings.position_embedding.scales": "model.safetensors",
761
  "vision_tower.vision_model.embeddings.position_embedding.weight": "model.safetensors",
762
  "vision_tower.vision_model.encoder.layers.0.layer_norm1.bias": "model.safetensors",
763
  "vision_tower.vision_model.encoder.layers.0.layer_norm1.weight": "model.safetensors",
tokenizer.json CHANGED
@@ -152,6 +152,7 @@
152
  "end_of_word_suffix": null,
153
  "fuse_unk": true,
154
  "byte_fallback": true,
 
155
  "vocab": {
156
  "<unk>": 0,
157
  "<s>": 1,
 
152
  "end_of_word_suffix": null,
153
  "fuse_unk": true,
154
  "byte_fallback": true,
155
+ "ignore_merges": false,
156
  "vocab": {
157
  "<unk>": 0,
158
  "<s>": 1,
tokenizer_config.json CHANGED
@@ -46,7 +46,6 @@
46
  "bos_token": "<s>",
47
  "clean_up_tokenization_spaces": false,
48
  "eos_token": "</s>",
49
- "legacy": false,
50
  "model_max_length": 1000000000000000019884624838656,
51
  "pad_token": "<pad>",
52
  "padding_side": "left",
 
46
  "bos_token": "<s>",
47
  "clean_up_tokenization_spaces": false,
48
  "eos_token": "</s>",
 
49
  "model_max_length": 1000000000000000019884624838656,
50
  "pad_token": "<pad>",
51
  "padding_side": "left",