Upload KawnIdefics3ForConditionalGeneration
Browse files- config.json +3 -2
- model-00001-of-00004.safetensors +2 -2
- model-00004-of-00004.safetensors +1 -1
- model.safetensors.index.json +1 -12
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"KawnIdefics3ForConditionalGeneration"
|
5 |
],
|
@@ -136,7 +136,8 @@
|
|
136 |
"model_type": "siglip_vision_model",
|
137 |
"num_attention_heads": 12,
|
138 |
"num_hidden_layers": 12,
|
139 |
-
"patch_size": 16
|
|
|
140 |
},
|
141 |
"vision_feature_layer": null
|
142 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "kawn_cohere_8b_idefics3_siglib14_384",
|
3 |
"architectures": [
|
4 |
"KawnIdefics3ForConditionalGeneration"
|
5 |
],
|
|
|
136 |
"model_type": "siglip_vision_model",
|
137 |
"num_attention_heads": 12,
|
138 |
"num_hidden_layers": 12,
|
139 |
+
"patch_size": 16,
|
140 |
+
"vision_use_head": false
|
141 |
},
|
142 |
"vision_feature_layer": null
|
143 |
}
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e3f2c2c9f817c2187e7fc3fc8e1f925ec4821e317a1097602bb0d0c88d76d874
|
3 |
+
size 4970053136
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1451273752
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:75850f91c5085aff924c8e090c475a003b941b7dded84d7bfb8bc10ea2b9379c
|
3 |
size 1451273752
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"connector.proj.weight": "model-00004-of-00004.safetensors",
|
@@ -457,17 +457,6 @@
|
|
457 |
"vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
458 |
"vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
459 |
"vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
460 |
-
"vision_tower.vision_model.head.attention.in_proj_bias": "model-00001-of-00004.safetensors",
|
461 |
-
"vision_tower.vision_model.head.attention.in_proj_weight": "model-00001-of-00004.safetensors",
|
462 |
-
"vision_tower.vision_model.head.attention.out_proj.bias": "model-00001-of-00004.safetensors",
|
463 |
-
"vision_tower.vision_model.head.attention.out_proj.weight": "model-00001-of-00004.safetensors",
|
464 |
-
"vision_tower.vision_model.head.layernorm.bias": "model-00001-of-00004.safetensors",
|
465 |
-
"vision_tower.vision_model.head.layernorm.weight": "model-00001-of-00004.safetensors",
|
466 |
-
"vision_tower.vision_model.head.mlp.fc1.bias": "model-00001-of-00004.safetensors",
|
467 |
-
"vision_tower.vision_model.head.mlp.fc1.weight": "model-00001-of-00004.safetensors",
|
468 |
-
"vision_tower.vision_model.head.mlp.fc2.bias": "model-00001-of-00004.safetensors",
|
469 |
-
"vision_tower.vision_model.head.mlp.fc2.weight": "model-00001-of-00004.safetensors",
|
470 |
-
"vision_tower.vision_model.head.probe": "model-00001-of-00004.safetensors",
|
471 |
"vision_tower.vision_model.post_layernorm.bias": "model-00001-of-00004.safetensors",
|
472 |
"vision_tower.vision_model.post_layernorm.weight": "model-00001-of-00004.safetensors"
|
473 |
}
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 16252918272
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"connector.proj.weight": "model-00004-of-00004.safetensors",
|
|
|
457 |
"vision_tower.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
|
458 |
"vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00004.safetensors",
|
459 |
"vision_tower.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
460 |
"vision_tower.vision_model.post_layernorm.bias": "model-00001-of-00004.safetensors",
|
461 |
"vision_tower.vision_model.post_layernorm.weight": "model-00001-of-00004.safetensors"
|
462 |
}
|