GuilhermeNunes commited on
Commit
d011549
·
verified ·
1 Parent(s): 925ec98

Upload LlavaNextForConditionalGeneration

Browse files
config.json CHANGED
@@ -66,7 +66,7 @@
66
  "multimodal_projector_bias": true,
67
  "projector_hidden_act": "gelu",
68
  "text_config": {
69
- "_name_or_path": "/mnt/scratch-artemis/gviveiros/TowerVision/llava-next-native/towerp_2b_base/",
70
  "add_faster_video": false,
71
  "add_time_instruction": false,
72
  "architectures": [
@@ -76,10 +76,12 @@
76
  "attention_dropout": 0.0,
77
  "attn_logit_softcapping": 50.0,
78
  "cache_implementation": "hybrid",
 
79
  "faster_token_stride": 10,
80
  "final_logit_softcapping": 30.0,
81
  "force_sample": false,
82
  "head_dim": 256,
 
83
  "hidden_activation": "gelu_pytorch_tanh",
84
  "hidden_size": 2304,
85
  "image_aspect_ratio": "anyres",
@@ -145,7 +147,7 @@
145
  "image_split_resolution": null,
146
  "initializer_range": 0.02,
147
  "intermediate_size": 9216,
148
- "max_position_embeddings": 4096,
149
  "mm_hidden_size": 1152,
150
  "mm_newline_position": "grid",
151
  "mm_patch_merge_type": "spatial_unpad",
 
66
  "multimodal_projector_bias": true,
67
  "projector_hidden_act": "gelu",
68
  "text_config": {
69
+ "_name_or_path": "/mnt/scratch-artemis/gviveiros/TowerVision/llava-next-native/towerp_2b_instruct/",
70
  "add_faster_video": false,
71
  "add_time_instruction": false,
72
  "architectures": [
 
76
  "attention_dropout": 0.0,
77
  "attn_logit_softcapping": 50.0,
78
  "cache_implementation": "hybrid",
79
+ "eos_token_id": 107,
80
  "faster_token_stride": 10,
81
  "final_logit_softcapping": 30.0,
82
  "force_sample": false,
83
  "head_dim": 256,
84
+ "hidden_act": "gelu_pytorch_tanh",
85
  "hidden_activation": "gelu_pytorch_tanh",
86
  "hidden_size": 2304,
87
  "image_aspect_ratio": "anyres",
 
147
  "image_split_resolution": null,
148
  "initializer_range": 0.02,
149
  "intermediate_size": 9216,
150
+ "max_position_embeddings": 8192,
151
  "mm_hidden_size": 1152,
152
  "mm_newline_position": "grid",
153
  "mm_patch_merge_type": "spatial_unpad",
generation_config.json CHANGED
@@ -2,7 +2,7 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 2,
4
  "cache_implementation": "hybrid",
5
- "eos_token_id": 1,
6
  "pad_token_id": 0,
7
  "transformers_version": "4.52.0.dev0"
8
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 2,
4
  "cache_implementation": "hybrid",
5
+ "eos_token_id": 107,
6
  "pad_token_id": 0,
7
  "transformers_version": "4.52.0.dev0"
8
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a337ac8a8db3caab6b1b452762056b731a5a057f3121ab0928244e1d08722246
3
  size 4978392440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a50290cb17c028cfd9e69030c9395071cd80d8ba84faba832bdfc6b0d2149aef
3
  size 4978392440
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aadce153b8abd872f2e01d65324a98621db80b32209c58df8a7c5ed8cb66de98
3
  size 1061826432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ab4b06305e64535bf8f944262d881a888b2cff49cbda8dc4e4f0407080ed7c3
3
  size 1061826432