PerRing commited on
Commit
df98191
1 Parent(s): 85a8af7

Upload LlavaForConditionalGeneration

Browse files
README.md CHANGED
@@ -1,3 +1,6 @@
 
 
 
1
  this repo is huggingface version of liuhaotian/llava-v1.6-34b
2
  # Issue
3
  Despite the completion of generation, '\n' is repeatedly generated, so be mindful of adjusting the 'max_length'.
 
1
+ ---
2
+ {}
3
+ ---
4
  this repo is huggingface version of liuhaotian/llava-v1.6-34b
5
  # Issue
6
  Despite the completion of generation, '\n' is repeatedly generated, so be mindful of adjusting the 'max_length'.
config.json CHANGED
@@ -1,17 +1,18 @@
1
  {
2
- "_name_or_path": "/backup/hgyoo/llava-v1.6-34b-hf",
3
  "architectures": [
4
  "LlavaForConditionalGeneration"
5
  ],
6
  "ignore_index": -100,
7
- "image_token_index": 64000,
8
  "model_type": "llava",
 
9
  "projector_hidden_act": "gelu",
10
  "text_config": {
11
- "_name_or_path": "liuhaotian/llava-v1.6-34b",
12
  "architectures": [
13
- "LlavaLlamaForCausalLM"
14
  ],
 
15
  "hidden_size": 7168,
16
  "intermediate_size": 20480,
17
  "max_position_embeddings": 4096,
@@ -19,9 +20,12 @@
19
  "num_attention_heads": 56,
20
  "num_hidden_layers": 60,
21
  "num_key_value_heads": 8,
 
22
  "rms_norm_eps": 1e-05,
 
23
  "torch_dtype": "bfloat16",
24
- "vocab_size": 64001
 
25
  },
26
  "torch_dtype": "float16",
27
  "transformers_version": "4.37.2",
@@ -38,5 +42,5 @@
38
  },
39
  "vision_feature_layer": -2,
40
  "vision_feature_select_strategy": "default",
41
- "vocab_size": 64001
42
  }
 
1
  {
 
2
  "architectures": [
3
  "LlavaForConditionalGeneration"
4
  ],
5
  "ignore_index": -100,
6
+ "image_token_index": 32000,
7
  "model_type": "llava",
8
+ "pad_token_id": 64001,
9
  "projector_hidden_act": "gelu",
10
  "text_config": {
11
+ "_name_or_path": "NousResearch/Nous-Hermes-2-Yi-34B",
12
  "architectures": [
13
+ "LlamaForCausalLM"
14
  ],
15
+ "eos_token_id": 7,
16
  "hidden_size": 7168,
17
  "intermediate_size": 20480,
18
  "max_position_embeddings": 4096,
 
20
  "num_attention_heads": 56,
21
  "num_hidden_layers": 60,
22
  "num_key_value_heads": 8,
23
+ "pad_token_id": 0,
24
  "rms_norm_eps": 1e-05,
25
+ "rope_theta": 5000000.0,
26
  "torch_dtype": "bfloat16",
27
+ "use_cache": false,
28
+ "vocab_size": 64064
29
  },
30
  "torch_dtype": "float16",
31
  "transformers_version": "4.37.2",
 
42
  },
43
  "vision_feature_layer": -2,
44
  "vision_feature_select_strategy": "default",
45
+ "vocab_size": 64064
46
  }
generation_config.json CHANGED
@@ -1,6 +1,8 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
- "eos_token_id": 2,
5
- "transformers_version": "4.37.2"
 
 
6
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
+ "eos_token_id": 7,
5
+ "pad_token_id": 64001,
6
+ "transformers_version": "4.37.2",
7
+ "use_cache": false
8
  }
model-00001-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69dbd39a3b5f302733a351967a3e8e1dd046b64a705ba8f47412a8b2fb21bf66
3
+ size 4990103872
model-00002-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c82a0bade77d76250cdf6fa6c9fb3effe865ef5864f6971a1b195269a356138
3
+ size 4991341784
model-00003-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bb6aa66b01e3e1abf211e631e0c3b9dbe7dc0d46ec76dda9f9cb3895e2a8907
3
+ size 4756460248
model-00004-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab7b1fe0e014c9b9e459e6be7ffbdd75ce8920853b14bbec71607e9bd3acb197
3
+ size 4991370760
model-00005-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19e8ffe65ae97578d739ceb9552f9ac21a56538d6be2111cc8fda48f4dd9e4f5
3
+ size 4756460280
model-00006-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01acb0fa038b6e6563b005fb7b23a3109929afd76e0d0d9d06ef9756c6ffa2d8
3
+ size 4756460272
model-00007-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a056b67345357b4b984402e06e65e8b3587b6603e7148f15c9e219770e84ee9
3
+ size 4991370760
model-00008-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:674c89b3196331ba4e08b399da3d5bf9df223ab08d89ee65ab47d0aef8fef563
3
+ size 4756460280
model-00009-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62b942f54a9cb906fd4b339b96eee97e5570f42415e21f060999a35f1ce063bb
3
+ size 4756460272
model-00010-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28f1803061fb0410196eb93d6f6b0118e56b9fda624d1f7c980f62351e56d3d7
3
+ size 4991370760
model-00011-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a74c9e5959d25022ec62850f802c82cc66807fbf838b3cb30323d598645c1c59
3
+ size 4756460280
model-00012-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd9d7518c977242ee857acfe57f2af3491b59ab7ef4012a51cfd534e7a1e8143
3
+ size 4756460272
model-00013-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7116bcb0da8bd73102c014277d86b7e14103037b42f6b8388116df05a4a8c3f2
3
+ size 4991370760
model-00014-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a12c6280bdd3dd3f735750e40e110a40bc730c18d1c063a70b9bd1e98acc5bb
3
+ size 4756460280
model-00015-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fc6bcc8a1ab2b4d4323f324f8b6ea5afad471f62049379b2a2434fe42947c3b
3
+ size 1505667832
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff