Xenova HF staff commited on
Commit
439caf3
1 Parent(s): f48018e

Upload PaliGemmaForConditionalGeneration

Browse files
Files changed (2) hide show
  1. config.json +4 -4
  2. model.safetensors +2 -2
config.json CHANGED
@@ -6,10 +6,10 @@
6
  "ignore_index": -100,
7
  "image_token_index": 257152,
8
  "model_type": "paligemma",
9
- "projection_dim": 16,
10
  "text_config": {
11
  "head_dim": 16,
12
- "hidden_size": 16,
13
  "intermediate_size": 32,
14
  "model_type": "gemma",
15
  "num_attention_heads": 4,
@@ -21,13 +21,13 @@
21
  "torch_dtype": "float32",
22
  "transformers_version": "4.42.0.dev0",
23
  "vision_config": {
24
- "hidden_size": 16,
25
  "intermediate_size": 32,
26
  "model_type": "siglip_vision_model",
27
  "num_attention_heads": 4,
28
  "num_hidden_layers": 2,
29
  "patch_size": 16,
30
- "projection_dim": 16
31
  },
32
  "vocab_size": 257152
33
  }
 
6
  "ignore_index": -100,
7
  "image_token_index": 257152,
8
  "model_type": "paligemma",
9
+ "projection_dim": 32,
10
  "text_config": {
11
  "head_dim": 16,
12
+ "hidden_size": 32,
13
  "intermediate_size": 32,
14
  "model_type": "gemma",
15
  "num_attention_heads": 4,
 
21
  "torch_dtype": "float32",
22
  "transformers_version": "4.42.0.dev0",
23
  "vision_config": {
24
+ "hidden_size": 32,
25
  "intermediate_size": 32,
26
  "model_type": "siglip_vision_model",
27
  "num_attention_heads": 4,
28
  "num_hidden_layers": 2,
29
  "patch_size": 16,
30
+ "projection_dim": 32
31
  },
32
  "vocab_size": 257152
33
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:662b66f05525622810c647d3a3e9e9b331c9dce7b3d85912c53acc293898d63b
3
- size 16605496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2aff118cf16a794a2831fd73c9fb7924e9489e37a1bb9aa40eee474c9e05c65
3
+ size 33228536