Upload new model version
Browse files- config.json +3 -8
- model-00001-of-00003.safetensors +1 -1
- model-00003-of-00003.safetensors +1 -1
- preprocessor_config.json +1 -0
- processor_config.json +2 -2
- tokenizer_config.json +1 -0
config.json
CHANGED
@@ -1,7 +1,4 @@
|
|
1 |
{
|
2 |
-
"architectures": [
|
3 |
-
"LlavaNextForConditionalGeneration"
|
4 |
-
],
|
5 |
"ignore_index": -100,
|
6 |
"image_grid_pinpoints": [
|
7 |
[
|
@@ -39,12 +36,11 @@
|
|
39 |
"model_type": "llama",
|
40 |
"pad_token_id": 0,
|
41 |
"rms_norm_eps": 1e-05,
|
42 |
-
"torch_dtype": "
|
43 |
"vocab_size": 32064
|
44 |
},
|
45 |
"tie_word_embeddings": false,
|
46 |
-
"
|
47 |
-
"transformers_version": "4.45.1",
|
48 |
"vision_config": {
|
49 |
"hidden_act": "gelu_pytorch_tanh",
|
50 |
"hidden_size": 1152,
|
@@ -54,8 +50,7 @@
|
|
54 |
"model_type": "siglip_vision_model",
|
55 |
"num_attention_heads": 16,
|
56 |
"num_hidden_layers": 27,
|
57 |
-
"patch_size": 14
|
58 |
-
"torch_dtype": "bfloat16"
|
59 |
},
|
60 |
"vision_feature_layer": -2,
|
61 |
"vision_feature_select_strategy": "full",
|
|
|
1 |
{
|
|
|
|
|
|
|
2 |
"ignore_index": -100,
|
3 |
"image_grid_pinpoints": [
|
4 |
[
|
|
|
36 |
"model_type": "llama",
|
37 |
"pad_token_id": 0,
|
38 |
"rms_norm_eps": 1e-05,
|
39 |
+
"torch_dtype": "float16",
|
40 |
"vocab_size": 32064
|
41 |
},
|
42 |
"tie_word_embeddings": false,
|
43 |
+
"transformers_version": "4.45.2",
|
|
|
44 |
"vision_config": {
|
45 |
"hidden_act": "gelu_pytorch_tanh",
|
46 |
"hidden_size": 1152,
|
|
|
50 |
"model_type": "siglip_vision_model",
|
51 |
"num_attention_heads": 16,
|
52 |
"num_hidden_layers": 27,
|
53 |
+
"patch_size": 14
|
|
|
54 |
},
|
55 |
"vision_feature_layer": -2,
|
56 |
"vision_feature_select_strategy": "full",
|
model-00001-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 9878580768
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e99cb04bec7aa0b851badca2f9516fc009e0746c40a4b70eedbe47ad694ebfa
|
3 |
size 9878580768
|
model-00003-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 8981442800
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02109df913624020c1c94c5aadbb06a665c02f038703fd1d004ed11c8ef5672b
|
3 |
size 8981442800
|
preprocessor_config.json
CHANGED
@@ -42,6 +42,7 @@
|
|
42 |
0.5,
|
43 |
0.5
|
44 |
],
|
|
|
45 |
"resample": 3,
|
46 |
"rescale_factor": 0.00392156862745098,
|
47 |
"size": {
|
|
|
42 |
0.5,
|
43 |
0.5
|
44 |
],
|
45 |
+
"processor_class": "LlavaNextProcessor",
|
46 |
"resample": 3,
|
47 |
"rescale_factor": 0.00392156862745098,
|
48 |
"size": {
|
processor_config.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"image_token": "<image>",
|
3 |
-
"patch_size":
|
4 |
"processor_class": "LlavaNextProcessor",
|
5 |
-
"vision_feature_select_strategy":
|
6 |
}
|
|
|
1 |
{
|
2 |
"image_token": "<image>",
|
3 |
+
"patch_size": 14,
|
4 |
"processor_class": "LlavaNextProcessor",
|
5 |
+
"vision_feature_select_strategy": "default"
|
6 |
}
|
tokenizer_config.json
CHANGED
@@ -37,6 +37,7 @@
|
|
37 |
}
|
38 |
},
|
39 |
"bos_token": "<s>",
|
|
|
40 |
"clean_up_tokenization_spaces": false,
|
41 |
"eos_token": "</s>",
|
42 |
"legacy": false,
|
|
|
37 |
}
|
38 |
},
|
39 |
"bos_token": "<s>",
|
40 |
+
"chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}",
|
41 |
"clean_up_tokenization_spaces": false,
|
42 |
"eos_token": "</s>",
|
43 |
"legacy": false,
|