Hennara commited on
Commit
aa29d97
·
verified ·
1 Parent(s): 7b03c78

Upload KawnIdefics3ForConditionalGeneration

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "kawn_idefics3_siglib14_384",
3
  "architectures": [
4
  "KawnIdefics3ForConditionalGeneration"
5
  ],
@@ -136,7 +136,8 @@
136
  "model_type": "siglip_vision_model",
137
  "num_attention_heads": 12,
138
  "num_hidden_layers": 12,
139
- "patch_size": 16
 
140
  },
141
  "vision_feature_layer": null
142
  }
 
1
  {
2
+ "_name_or_path": "kawn_cohere_8b_idefics3_siglib14_384",
3
  "architectures": [
4
  "KawnIdefics3ForConditionalGeneration"
5
  ],
 
136
  "model_type": "siglip_vision_model",
137
  "num_attention_heads": 12,
138
  "num_hidden_layers": 12,
139
+ "patch_size": 16,
140
+ "vision_use_head": false
141
  },
142
  "vision_feature_layer": null
143
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b310fd907084c3846721f58067147e1c6a35670b1edcf2492bc794db83005f2d
3
- size 4984228696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3f2c2c9f817c2187e7fc3fc8e1f925ec4821e317a1097602bb0d0c88d76d874
3
+ size 4970053136
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a7b843a130374112910730baa71ff3784aa650e995a63c810c4395b0e93b744
3
  size 1451273752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75850f91c5085aff924c8e090c475a003b941b7dded84d7bfb8bc10ea2b9379c
3
  size 1451273752
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 16267092480
4
  },
5
  "weight_map": {
6
  "connector.proj.weight": "model-00004-of-00004.safetensors",
@@ -457,17 +457,6 @@
457
  "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
458
  "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
459
  "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
460
- "vision_tower.vision_model.head.attention.in_proj_bias": "model-00001-of-00004.safetensors",
461
- "vision_tower.vision_model.head.attention.in_proj_weight": "model-00001-of-00004.safetensors",
462
- "vision_tower.vision_model.head.attention.out_proj.bias": "model-00001-of-00004.safetensors",
463
- "vision_tower.vision_model.head.attention.out_proj.weight": "model-00001-of-00004.safetensors",
464
- "vision_tower.vision_model.head.layernorm.bias": "model-00001-of-00004.safetensors",
465
- "vision_tower.vision_model.head.layernorm.weight": "model-00001-of-00004.safetensors",
466
- "vision_tower.vision_model.head.mlp.fc1.bias": "model-00001-of-00004.safetensors",
467
- "vision_tower.vision_model.head.mlp.fc1.weight": "model-00001-of-00004.safetensors",
468
- "vision_tower.vision_model.head.mlp.fc2.bias": "model-00001-of-00004.safetensors",
469
- "vision_tower.vision_model.head.mlp.fc2.weight": "model-00001-of-00004.safetensors",
470
- "vision_tower.vision_model.head.probe": "model-00001-of-00004.safetensors",
471
  "vision_tower.vision_model.post_layernorm.bias": "model-00001-of-00004.safetensors",
472
  "vision_tower.vision_model.post_layernorm.weight": "model-00001-of-00004.safetensors"
473
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 16252918272
4
  },
5
  "weight_map": {
6
  "connector.proj.weight": "model-00004-of-00004.safetensors",
 
457
  "vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
458
  "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
459
  "vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
 
 
 
 
 
 
 
 
 
 
 
460
  "vision_tower.vision_model.post_layernorm.bias": "model-00001-of-00004.safetensors",
461
  "vision_tower.vision_model.post_layernorm.weight": "model-00001-of-00004.safetensors"
462
  }