yujiepan commited on
Commit
c1118af
1 Parent(s): ae7948f

Upload folder using huggingface_hub

Browse files
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "</s>": 2,
3
+ "<pad>": 1
4
+ }
compressed_graph.dot ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/opt-6.7b",
3
+ "_remove_final_layer_norm": false,
4
+ "activation_dropout": 0.0,
5
+ "activation_function": "relu",
6
+ "architectures": [
7
+ "OPTForCausalLM"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "bos_token_id": 2,
11
+ "do_layer_norm_before": true,
12
+ "dropout": 0.1,
13
+ "enable_bias": true,
14
+ "eos_token_id": 2,
15
+ "ffn_dim": 16384,
16
+ "hidden_size": 4096,
17
+ "init_std": 0.02,
18
+ "layer_norm_elementwise_affine": true,
19
+ "layerdrop": 0.0,
20
+ "max_position_embeddings": 2048,
21
+ "model_type": "opt",
22
+ "num_attention_heads": 32,
23
+ "num_hidden_layers": 32,
24
+ "pad_token_id": 1,
25
+ "prefix": "</s>",
26
+ "torch_dtype": "float16",
27
+ "transformers_version": "4.34.0",
28
+ "use_cache": true,
29
+ "vocab_size": 50272,
30
+ "word_embed_proj_dim": 4096
31
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
openvino_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "compression": {
3
+ "algorithm": "quantization",
4
+ "export_to_onnx_standard_ops": false,
5
+ "ignored_scopes": [
6
+ "{re}.*Embedding.*",
7
+ "{re}.*add___.*",
8
+ "{re}.*layer_norm_.*"
9
+ ],
10
+ "initializer": {
11
+ "batchnorm_adaptation": {
12
+ "num_bn_adaptation_samples": 0
13
+ },
14
+ "range": {
15
+ "num_init_samples": 16,
16
+ "type": "min_max"
17
+ }
18
+ },
19
+ "overflow_fix": "disable",
20
+ "preset": "performance"
21
+ },
22
+ "input_info": [
23
+ {
24
+ "keyword": "input_ids",
25
+ "sample_size": [
26
+ 1,
27
+ 32
28
+ ],
29
+ "type": "long"
30
+ },
31
+ {
32
+ "keyword": "attention_mask",
33
+ "sample_size": [
34
+ 1,
35
+ 32
36
+ ],
37
+ "type": "long"
38
+ }
39
+ ],
40
+ "log_dir": "./logs/opt-6.7b-w8w8-unstructured50/",
41
+ "optimum_version": "1.13.2",
42
+ "save_onnx_model": false,
43
+ "transformers_version": "4.34.0"
44
+ }
openvino_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78c59d0a9a9349be156fbe0cef16f776555f2a1293ffdf200be1579faa287468
3
+ size 7516461328
openvino_model.xml ADDED
The diff for this file is too large to render. See raw diff
 
original_graph.dot ADDED
The diff for this file is too large to render. See raw diff
 
ov_sparsity_stats.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "Constant_740364": 0.5002646446228027,
3
+ "Constant_740368": 0.5002545714378357,
4
+ "Constant_740372": 0.5001242160797119,
5
+ "Constant_740376": 0.5001725554466248,
6
+ "Constant_740380": 0.5000810772180557,
7
+ "Constant_740384": 0.50013767182827,
8
+ "Constant_740388": 0.5000916123390198,
9
+ "Constant_740392": 0.5001347064971924,
10
+ "Constant_740396": 0.5001572370529175,
11
+ "Constant_740400": 0.5001401305198669,
12
+ "Constant_740404": 0.5002889931201935,
13
+ "Constant_740408": 0.5001097619533539,
14
+ "Constant_740412": 0.5000396966934204,
15
+ "Constant_740416": 0.5000334978103638,
16
+ "Constant_740420": 0.500091016292572,
17
+ "Constant_740424": 0.5000038146972656,
18
+ "Constant_740428": 0.5001920312643051,
19
+ "Constant_740432": 0.5002889633178711,
20
+ "Constant_740436": 0.5001274347305298,
21
+ "Constant_740440": 0.5002991557121277,
22
+ "Constant_740444": 0.5000097155570984,
23
+ "Constant_740448": 0.5003159642219543,
24
+ "Constant_740452": 0.5002010613679886,
25
+ "Constant_740456": 0.5002553313970566,
26
+ "Constant_740460": 0.5001412630081177,
27
+ "Constant_740464": 0.5001715421676636,
28
+ "Constant_740468": 0.5000601410865784,
29
+ "Constant_740472": 0.5000429749488831,
30
+ "Constant_740476": 0.5002230852842331,
31
+ "Constant_740480": 0.5002836287021637,
32
+ "Constant_740484": 0.5001240372657776,
33
+ "Constant_740488": 0.5000519752502441,
34
+ "Constant_740492": 0.5000098347663879,
35
+ "Constant_740496": 0.5000232458114624,
36
+ "Constant_740500": 0.5000854283571243,
37
+ "Constant_740504": 0.5000220388174057,
38
+ "Constant_740508": 0.5000513792037964,
39
+ "Constant_740512": 0.5000736713409424,
40
+ "Constant_740516": 0.5001561045646667,
41
+ "Constant_740520": 0.5001461505889893,
42
+ "Constant_740524": 0.5000685006380081,
43
+ "Constant_740528": 0.5001946538686752,
44
+ "Constant_740532": 0.5001165866851807,
45
+ "Constant_740536": 0.5002720952033997,
46
+ "Constant_740540": 0.5002614259719849,
47
+ "Constant_740544": 0.5001602172851562,
48
+ "Constant_740548": 0.5002289712429047,
49
+ "Constant_740552": 0.5001602917909622,
50
+ "Constant_740556": 0.5000261068344116,
51
+ "Constant_740560": 0.500288188457489,
52
+ "Constant_740564": 0.5002854466438293,
53
+ "Constant_740568": 0.5000322461128235,
54
+ "Constant_740572": 0.5001158714294434,
55
+ "Constant_740576": 0.5001968443393707,
56
+ "Constant_740580": 0.5000325441360474,
57
+ "Constant_740584": 0.500220775604248,
58
+ "Constant_740588": 0.5002447366714478,
59
+ "Constant_740592": 0.5002323389053345,
60
+ "Constant_740596": 0.5001165121793747,
61
+ "Constant_740600": 0.5001854747533798,
62
+ "Constant_740604": 0.5000985264778137,
63
+ "Constant_740608": 0.5000840425491333,
64
+ "Constant_740612": 0.5001408457756042,
65
+ "Constant_740616": 0.5001469254493713,
66
+ "Constant_740620": 0.5002485513687134,
67
+ "Constant_740624": 0.5000925958156586,
68
+ "Constant_740628": 0.5001927614212036,
69
+ "Constant_740632": 0.5001641511917114,
70
+ "Constant_740636": 0.5002038478851318,
71
+ "Constant_740640": 0.5001091361045837,
72
+ "Constant_740644": 0.5001162588596344,
73
+ "Constant_740648": 0.5002866089344025,
74
+ "Constant_740652": 0.5001062154769897,
75
+ "Constant_740656": 0.5000807642936707,
76
+ "Constant_740660": 0.5003570318222046,
77
+ "Constant_740664": 0.5001456141471863,
78
+ "Constant_740668": 0.5002349317073822,
79
+ "Constant_740672": 0.500250443816185,
80
+ "Constant_740676": 0.5001181364059448,
81
+ "Constant_740680": 0.5002414584159851,
82
+ "Constant_740684": 0.5002016425132751,
83
+ "Constant_740688": 0.5001419186592102,
84
+ "Constant_740692": 0.5002192854881287,
85
+ "Constant_740696": 0.5002044141292572,
86
+ "Constant_740700": 0.5001293420791626,
87
+ "Constant_740704": 0.5000990629196167,
88
+ "Constant_740708": 0.5001548528671265,
89
+ "Constant_740712": 0.5000123381614685,
90
+ "Constant_740716": 0.5000330209732056,
91
+ "Constant_740720": 0.5000393390655518,
92
+ "Constant_740724": 0.5000688433647156,
93
+ "Constant_740728": 0.5001242160797119,
94
+ "Constant_740732": 0.5002279877662659,
95
+ "Constant_740736": 0.5000247359275818,
96
+ "Constant_740740": 0.500240683555603,
97
+ "Constant_740744": 0.5000388026237488,
98
+ "Constant_740748": 0.5001137852668762,
99
+ "Constant_740752": 0.5000342726707458,
100
+ "Constant_740756": 0.5001882314682007,
101
+ "Constant_740760": 0.5003361105918884,
102
+ "Constant_740764": 0.5002809017896652,
103
+ "Constant_740768": 0.5002386122941971,
104
+ "Constant_740772": 0.500178337097168,
105
+ "Constant_740776": 0.5001136660575867,
106
+ "Constant_740780": 0.5001077651977539,
107
+ "Constant_740784": 0.5003582239151001,
108
+ "Constant_740788": 0.5000877976417542,
109
+ "Constant_740792": 0.5002572685480118,
110
+ "Constant_740796": 0.5001864433288574,
111
+ "Constant_740800": 0.5001525282859802,
112
+ "Constant_740804": 0.5002452731132507,
113
+ "Constant_740808": 0.5003000497817993,
114
+ "Constant_740812": 0.500264897942543,
115
+ "Constant_740816": 0.5001865923404694,
116
+ "Constant_740820": 0.5002719759941101,
117
+ "Constant_740824": 0.5002580285072327,
118
+ "Constant_740828": 0.5001961588859558,
119
+ "Constant_740832": 0.5003326535224915,
120
+ "Constant_740836": 0.5000276118516922,
121
+ "Constant_740840": 0.500083327293396,
122
+ "Constant_740844": 0.5002673864364624,
123
+ "Constant_740848": 0.5000994801521301,
124
+ "Constant_740852": 0.5002595782279968,
125
+ "Constant_740856": 0.5001588463783264,
126
+ "Constant_740860": 0.5000343769788742,
127
+ "Constant_740864": 0.5002723783254623,
128
+ "Constant_740868": 0.5001829266548157,
129
+ "Constant_740872": 0.5002307295799255,
130
+ "Constant_740876": 0.5000202059745789,
131
+ "Constant_740880": 0.5001521110534668,
132
+ "Constant_740884": 0.5001227259635925,
133
+ "Constant_740888": 0.5002606809139252,
134
+ "Constant_740892": 0.5002914667129517,
135
+ "Constant_740896": 0.5001171827316284,
136
+ "Constant_740900": 0.5000696182250977,
137
+ "Constant_740904": 0.5002221465110779,
138
+ "Constant_740908": 0.5001393109560013,
139
+ "Constant_740912": 0.5002193599939346,
140
+ "Constant_740916": 0.5000544786453247,
141
+ "Constant_740920": 0.5001165866851807,
142
+ "Constant_740924": 0.5000243186950684,
143
+ "Constant_740928": 0.5000439286231995,
144
+ "Constant_740932": 0.5000893771648407,
145
+ "Constant_740936": 0.500079557299614,
146
+ "Constant_740940": 0.5002276301383972,
147
+ "Constant_740944": 0.500243067741394,
148
+ "Constant_740948": 0.5001072287559509,
149
+ "Constant_740952": 0.5000651478767395,
150
+ "Constant_740956": 0.5001714676618576,
151
+ "Constant_740960": 0.5000118166208267,
152
+ "Constant_740964": 0.5000373721122742,
153
+ "Constant_740968": 0.5002548694610596,
154
+ "Constant_740972": 0.5001973509788513,
155
+ "Constant_740976": 0.5001676082611084,
156
+ "Constant_740980": 0.5000311583280563,
157
+ "Constant_740984": 0.5000687539577484,
158
+ "Constant_740988": 0.5003244876861572,
159
+ "Constant_740992": 0.5001760125160217,
160
+ "Constant_740996": 0.5001277923583984,
161
+ "Constant_741000": 0.5002439618110657,
162
+ "Constant_741004": 0.5001821517944336,
163
+ "Constant_741008": 0.5001326352357864,
164
+ "Constant_741012": 0.5003024935722351,
165
+ "Constant_741016": 0.5000227093696594,
166
+ "Constant_741020": 0.5000532865524292,
167
+ "Constant_741024": 0.5000585913658142,
168
+ "Constant_741028": 0.500203400850296,
169
+ "Constant_741032": 0.500085711479187,
170
+ "Constant_741036": 0.5000766515731812,
171
+ "Constant_741040": 0.5001662373542786,
172
+ "Constant_741044": 0.5001732707023621,
173
+ "Constant_741048": 0.5001737475395203,
174
+ "Constant_741052": 0.5000820159912109,
175
+ "Constant_741056": 0.5001400411128998,
176
+ "Constant_741060": 0.5001081824302673,
177
+ "Constant_741064": 0.5001533627510071,
178
+ "Constant_741068": 0.5001651048660278,
179
+ "Constant_741072": 0.5000892877578735,
180
+ "Constant_741076": 0.5000907778739929,
181
+ "Constant_741080": 0.5000842660665512,
182
+ "Constant_741084": 0.5001571774482727,
183
+ "Constant_741088": 0.5000737309455872,
184
+ "Constant_741092": 0.5000970959663391,
185
+ "Constant_741096": 0.5000082850456238,
186
+ "Constant_741100": 0.5002415478229523,
187
+ "Constant_741104": 0.5001385509967804,
188
+ "Constant_741108": 0.5000050067901611,
189
+ "Constant_741112": 0.5001327395439148,
190
+ "Constant_741116": 0.500109076499939,
191
+ "Constant_741120": 0.500032901763916,
192
+ "Constant_741124": 0.5000126212835312,
193
+ "Constant_741128": 0.5000438243150711,
194
+ "Constant_741132": 0.013297121665949733
195
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "</s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "</s>",
5
+ "unk_token": "</s>"
6
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "1": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "2": {
14
+ "content": "</s>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ }
21
+ },
22
+ "additional_special_tokens": [],
23
+ "bos_token": "</s>",
24
+ "clean_up_tokenization_spaces": true,
25
+ "eos_token": "</s>",
26
+ "errors": "replace",
27
+ "model_max_length": 1000000000000000019884624838656,
28
+ "pad_token": "</s>",
29
+ "tokenizer_class": "GPT2Tokenizer",
30
+ "unk_token": "</s>"
31
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff