musashihinck commited on
Commit
7de365f
1 Parent(s): a2ed3a6

Converting weights

Browse files
README.md CHANGED
@@ -1,8 +1,8 @@
1
  ---
2
- license_name: gemma-terms
3
- license_link: https://ai.google.dev/gemma/terms
4
  language:
5
  - en
 
 
6
  ---
7
 
8
  # LLaVA-Gemma Model Card
 
1
  ---
 
 
2
  language:
3
  - en
4
+ license_name: gemma-terms
5
+ license_link: https://ai.google.dev/gemma/terms
6
  ---
7
 
8
  # LLaVA-Gemma Model Card
config.json CHANGED
@@ -1,43 +1,48 @@
1
  {
2
- "_name_or_path": "/root/.cache/huggingface/hub/models--google--gemma-2b-it/snapshots/1027d96c1638a27f01ae935cd98bac7d1a01686c/",
3
  "architectures": [
4
- "LlavaGemmaForCausalLM"
5
  ],
6
- "attention_bias": false,
7
- "attention_dropout": 0.0,
8
- "bos_token_id": 2,
9
- "eos_token_id": 1,
10
- "freeze_mm_mlp_adapter": false,
11
- "head_dim": 256,
12
- "hidden_act": "gelu",
13
- "hidden_size": 2048,
14
- "image_aspect_ratio": "pad",
15
- "initializer_range": 0.02,
16
- "intermediate_size": 16384,
17
- "max_position_embeddings": 8192,
18
- "mm_hidden_size": 1024,
19
- "mm_patch_merge_type": "flat",
20
- "mm_projector_lr": null,
21
- "mm_projector_type": "mlp2x_gelu",
22
- "mm_use_im_patch_token": false,
23
- "mm_use_im_start_end": false,
24
- "mm_vision_select_feature": "patch",
25
- "mm_vision_select_layer": -2,
26
- "mm_vision_tower": "openai/clip-vit-large-patch14-336",
27
- "model_type": "llava_gemma",
28
- "num_attention_heads": 8,
29
- "num_hidden_layers": 18,
30
- "num_key_value_heads": 1,
31
- "pad_token_id": 0,
32
- "rms_norm_eps": 1e-06,
33
- "rope_scaling": null,
34
- "rope_theta": 10000.0,
35
- "tokenizer_model_max_length": 2048,
36
- "tokenizer_padding_side": "right",
37
- "torch_dtype": "bfloat16",
38
  "transformers_version": "4.39.0.dev0",
39
- "tune_mm_mlp_adapter": false,
40
- "use_cache": false,
41
- "use_mm_proj": true,
42
- "vocab_size": 256000
 
 
 
 
 
 
 
 
 
 
43
  }
 
1
  {
2
+ "_name_or_path": "./llava-gemma-2b-it",
3
  "architectures": [
4
+ "LlavaForConditionalGeneration"
5
  ],
6
+ "ignore_index": -100,
7
+ "image_token_index": 256000,
8
+ "model_type": "llava",
9
+ "projector_hidden_act": "gelu",
10
+ "text_config": {
11
+ "_name_or_path": "google/gemma-2b-it",
12
+ "architectures": [
13
+ "GemmaForCausalLM"
14
+ ],
15
+ "bos_token_id": 2,
16
+ "eos_token_id": 1,
17
+ "head_dim": 256,
18
+ "hidden_act": "gelu",
19
+ "hidden_size": 2048,
20
+ "intermediate_size": 16384,
21
+ "max_position_embeddings": 8192,
22
+ "model_type": "gemma",
23
+ "num_attention_heads": 8,
24
+ "num_hidden_layers": 18,
25
+ "num_key_value_heads": 1,
26
+ "pad_token_id": 0,
27
+ "rope_scaling": null,
28
+ "tie_word_embeddings": true,
29
+ "torch_dtype": "bfloat16",
30
+ "vocab_size": 256064
31
+ },
32
+ "torch_dtype": "float32",
 
 
 
 
 
33
  "transformers_version": "4.39.0.dev0",
34
+ "vision_config": {
35
+ "hidden_size": 1024,
36
+ "image_size": 336,
37
+ "intermediate_size": 4096,
38
+ "model_type": "clip_vision_model",
39
+ "num_attention_heads": 16,
40
+ "num_hidden_layers": 24,
41
+ "patch_size": 14,
42
+ "projection_dim": 768,
43
+ "vocab_size": 32000
44
+ },
45
+ "vision_feature_layer": -2,
46
+ "vision_feature_select_strategy": "default",
47
+ "vocab_size": 256064
48
  }
generation_config.json CHANGED
@@ -1,19 +1,7 @@
1
  {
2
  "_from_model_config": true,
3
- "attn_softmax_bf16": null,
4
  "bos_token_id": 2,
5
- "bucket_internal": null,
6
- "bucket_size": -1,
7
  "eos_token_id": 1,
8
- "flash_attention_recompute": null,
9
- "ignore_eos": null,
10
- "kv_cache_fp8": null,
11
- "limit_hpu_graphs": null,
12
  "pad_token_id": 0,
13
- "reduce_recompile": null,
14
- "reuse_cache": null,
15
- "static_shapes": null,
16
- "transformers_version": "4.39.0.dev0",
17
- "trim_logits": null,
18
- "use_flash_attention": null
19
  }
 
1
  {
2
  "_from_model_config": true,
 
3
  "bos_token_id": 2,
 
 
4
  "eos_token_id": 1,
 
 
 
 
5
  "pad_token_id": 0,
6
+ "transformers_version": "4.39.0.dev0"
 
 
 
 
 
7
  }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a34f002270e12d50917ff5d307951ed107140530372c2d549aea081e327b4f6f
3
+ size 4964385864
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21d7fe224a3dced4a95cd2316b9d131a5c80705115849dd980eafbfe108d5e5c
3
+ size 4999820616
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40c198b0d6b4170bd3d8eb708a66a672b76ae70eece7ab843b8ed3605d97c85c
3
+ size 1300294808
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff