Xenova HF staff commited on
Commit
ba08b91
1 Parent(s): 4207ed6

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_commit_hash": "dc68f91c06a1ba6f15268e5b9c13ae7a7c514084",
3
  "_name_or_path": "nlpconnect/vit-gpt2-image-captioning",
4
  "architectures": [
5
  "VisionEncoderDecoderModel"
@@ -92,7 +91,6 @@
92
  "top_p": 1.0,
93
  "torch_dtype": null,
94
  "torchscript": false,
95
- "transformers_version": "4.29.2",
96
  "typical_p": 1.0,
97
  "use_bfloat16": false,
98
  "use_cache": true,
@@ -175,7 +173,6 @@
175
  "top_p": 1.0,
176
  "torch_dtype": null,
177
  "torchscript": false,
178
- "transformers_version": "4.29.2",
179
  "typical_p": 1.0,
180
  "use_bfloat16": false
181
  },
@@ -184,5 +181,5 @@
184
  "model_type": "vision-encoder-decoder",
185
  "pad_token_id": 50256,
186
  "tie_word_embeddings": false,
187
- "transformers_version": null
188
  }
 
1
  {
 
2
  "_name_or_path": "nlpconnect/vit-gpt2-image-captioning",
3
  "architectures": [
4
  "VisionEncoderDecoderModel"
 
91
  "top_p": 1.0,
92
  "torch_dtype": null,
93
  "torchscript": false,
 
94
  "typical_p": 1.0,
95
  "use_bfloat16": false,
96
  "use_cache": true,
 
173
  "top_p": 1.0,
174
  "torch_dtype": null,
175
  "torchscript": false,
 
176
  "typical_p": 1.0,
177
  "use_bfloat16": false
178
  },
 
181
  "model_type": "vision-encoder-decoder",
182
  "pad_token_id": 50256,
183
  "tie_word_embeddings": false,
184
+ "transformers_version": "4.33.0.dev0"
185
  }
generation_config.json CHANGED
@@ -4,5 +4,5 @@
4
  "decoder_start_token_id": 50256,
5
  "eos_token_id": 50256,
6
  "pad_token_id": 50256,
7
- "transformers_version": "4.29.2"
8
  }
 
4
  "decoder_start_token_id": 50256,
5
  "eos_token_id": 50256,
6
  "pad_token_id": 50256,
7
+ "transformers_version": "4.33.0.dev0"
8
  }
onnx/decoder_model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:722fc4a49f257eb69381633d21107308d01914601facf7f2143e0dedc4f4348a
3
- size 767520869
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14c3b794c998e17cad6f5e510504a6ce4e9607fb82e130907d39c9a6deb8f310
3
+ size 613129810
onnx/decoder_model_merged.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:402ec6b9822f04e89fe325996d0aeb41d17ce239ce7828c85fb87093c10ebf6a
3
- size 768369266
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df4abab679e8fa8d5c2140dd1ce83d136e7d47bc08008c9d8b3860643c502776
3
+ size 615025088
onnx/decoder_model_merged_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c0f5f195147ca4087167acb7c9cd37b084bd12d1dea8792d23e0d0532630be7
3
- size 196406021
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f3ec53b5fc3614c0b54ee786755a7ec3007841f57543f8e241499edfadfa98f
3
+ size 158599996
onnx/decoder_model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b06523e3e4832e2aeec9aa7c57785e9792c56871a0cf82c5171445932fc9b881
3
- size 195129787
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cd3a527b1acd9893dec2ef8f02b5e759c0365e365976a2f44313c6095a8a256
3
+ size 156278186
onnx/decoder_with_past_model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6258fd83496643ae292beef05d3fe9e6968d425099d7353d34c678308d09bc98
3
- size 767518177
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:926a856e44e926f6a865e8225009752615e7903a4a0b5641ff5ee9fe65cb9f14
3
+ size 613127118
onnx/decoder_with_past_model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70af4a8110df4578662bb971ec10bcfdc045b7691ed7eb7e0ce51d4be8886ef8
3
- size 195121650
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a72d2c3557b4b5728e67eca3124682e0f3f1ed8aba1ae194556d03eaf3ccc693
3
+ size 156270049
quantize_config.json ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "per_channel": true,
3
+ "reduce_range": true,
4
+ "per_model_config": {
5
+ "decoder_model": {
6
+ "op_types": [
7
+ "Add",
8
+ "Where",
9
+ "MatMul",
10
+ "Gather",
11
+ "Gemm",
12
+ "Reshape",
13
+ "Concat",
14
+ "Cast",
15
+ "Pow",
16
+ "Constant",
17
+ "ConstantOfShape",
18
+ "Slice",
19
+ "Mul",
20
+ "Div",
21
+ "Unsqueeze",
22
+ "Sub",
23
+ "Squeeze",
24
+ "Softmax",
25
+ "Split",
26
+ "Transpose",
27
+ "Shape",
28
+ "Tanh",
29
+ "Range",
30
+ "ReduceMean",
31
+ "Sqrt"
32
+ ],
33
+ "weight_type": "QInt8"
34
+ },
35
+ "encoder_model": {
36
+ "op_types": [
37
+ "Where",
38
+ "Add",
39
+ "Equal",
40
+ "MatMul",
41
+ "Gather",
42
+ "Conv",
43
+ "Reshape",
44
+ "Concat",
45
+ "Expand",
46
+ "Pow",
47
+ "Constant",
48
+ "ConstantOfShape",
49
+ "Erf",
50
+ "Slice",
51
+ "Mul",
52
+ "Div",
53
+ "Unsqueeze",
54
+ "Sub",
55
+ "Softmax",
56
+ "Transpose",
57
+ "Shape",
58
+ "ReduceMean",
59
+ "Sqrt"
60
+ ],
61
+ "weight_type": "QUInt8"
62
+ },
63
+ "decoder_model_merged": {
64
+ "op_types": [
65
+ "Add",
66
+ "Where",
67
+ "MatMul",
68
+ "Gather",
69
+ "Gemm",
70
+ "Reshape",
71
+ "Concat",
72
+ "Cast",
73
+ "Pow",
74
+ "Constant",
75
+ "ConstantOfShape",
76
+ "Slice",
77
+ "Mul",
78
+ "If",
79
+ "Div",
80
+ "Unsqueeze",
81
+ "Sub",
82
+ "Squeeze",
83
+ "Softmax",
84
+ "Split",
85
+ "Transpose",
86
+ "Shape",
87
+ "Tanh",
88
+ "Range",
89
+ "ReduceMean",
90
+ "Sqrt"
91
+ ],
92
+ "weight_type": "QInt8"
93
+ },
94
+ "decoder_with_past_model": {
95
+ "op_types": [
96
+ "Add",
97
+ "Where",
98
+ "MatMul",
99
+ "Gather",
100
+ "Gemm",
101
+ "Reshape",
102
+ "Concat",
103
+ "Cast",
104
+ "Pow",
105
+ "Constant",
106
+ "ConstantOfShape",
107
+ "Slice",
108
+ "Mul",
109
+ "Div",
110
+ "Unsqueeze",
111
+ "Squeeze",
112
+ "Sub",
113
+ "Softmax",
114
+ "Split",
115
+ "Transpose",
116
+ "Shape",
117
+ "Tanh",
118
+ "Range",
119
+ "ReduceMean",
120
+ "Sqrt"
121
+ ],
122
+ "weight_type": "QInt8"
123
+ }
124
+ }
125
+ }
tokenizer_config.json CHANGED
@@ -3,7 +3,15 @@
3
  "bos_token": "<|endoftext|>",
4
  "clean_up_tokenization_spaces": true,
5
  "eos_token": "<|endoftext|>",
 
6
  "model_max_length": 1024,
 
 
 
 
 
7
  "tokenizer_class": "GPT2Tokenizer",
 
 
8
  "unk_token": "<|endoftext|>"
9
  }
 
3
  "bos_token": "<|endoftext|>",
4
  "clean_up_tokenization_spaces": true,
5
  "eos_token": "<|endoftext|>",
6
+ "max_length": 32,
7
  "model_max_length": 1024,
8
+ "pad_to_multiple_of": null,
9
+ "pad_token": "<|endoftext|>",
10
+ "pad_token_type_id": 0,
11
+ "padding_side": "right",
12
+ "stride": 0,
13
  "tokenizer_class": "GPT2Tokenizer",
14
+ "truncation_side": "right",
15
+ "truncation_strategy": "longest_first",
16
  "unk_token": "<|endoftext|>"
17
  }