Upload folder using huggingface_hub
Browse files- added_tokens.json +4 -0
- compressed_graph.dot +0 -0
- config.json +31 -0
- merges.txt +0 -0
- openvino_config.json +44 -0
- openvino_model.bin +3 -0
- openvino_model.xml +0 -0
- original_graph.dot +0 -0
- ov_sparsity_stats.json +195 -0
- special_tokens_map.json +6 -0
- tokenizer.json +0 -0
- tokenizer_config.json +31 -0
- vocab.json +0 -0
added_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"</s>": 2,
|
3 |
+
"<pad>": 1
|
4 |
+
}
|
compressed_graph.dot
ADDED
The diff for this file is too large to render.
See raw diff
|
|
config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "facebook/opt-6.7b",
|
3 |
+
"_remove_final_layer_norm": false,
|
4 |
+
"activation_dropout": 0.0,
|
5 |
+
"activation_function": "relu",
|
6 |
+
"architectures": [
|
7 |
+
"OPTForCausalLM"
|
8 |
+
],
|
9 |
+
"attention_dropout": 0.0,
|
10 |
+
"bos_token_id": 2,
|
11 |
+
"do_layer_norm_before": true,
|
12 |
+
"dropout": 0.1,
|
13 |
+
"enable_bias": true,
|
14 |
+
"eos_token_id": 2,
|
15 |
+
"ffn_dim": 16384,
|
16 |
+
"hidden_size": 4096,
|
17 |
+
"init_std": 0.02,
|
18 |
+
"layer_norm_elementwise_affine": true,
|
19 |
+
"layerdrop": 0.0,
|
20 |
+
"max_position_embeddings": 2048,
|
21 |
+
"model_type": "opt",
|
22 |
+
"num_attention_heads": 32,
|
23 |
+
"num_hidden_layers": 32,
|
24 |
+
"pad_token_id": 1,
|
25 |
+
"prefix": "</s>",
|
26 |
+
"torch_dtype": "float16",
|
27 |
+
"transformers_version": "4.34.0",
|
28 |
+
"use_cache": true,
|
29 |
+
"vocab_size": 50272,
|
30 |
+
"word_embed_proj_dim": 4096
|
31 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
openvino_config.json
ADDED
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"compression": {
|
3 |
+
"algorithm": "quantization",
|
4 |
+
"export_to_onnx_standard_ops": false,
|
5 |
+
"ignored_scopes": [
|
6 |
+
"{re}.*Embedding.*",
|
7 |
+
"{re}.*add___.*",
|
8 |
+
"{re}.*layer_norm_.*"
|
9 |
+
],
|
10 |
+
"initializer": {
|
11 |
+
"batchnorm_adaptation": {
|
12 |
+
"num_bn_adaptation_samples": 0
|
13 |
+
},
|
14 |
+
"range": {
|
15 |
+
"num_init_samples": 16,
|
16 |
+
"type": "min_max"
|
17 |
+
}
|
18 |
+
},
|
19 |
+
"overflow_fix": "disable",
|
20 |
+
"preset": "performance"
|
21 |
+
},
|
22 |
+
"input_info": [
|
23 |
+
{
|
24 |
+
"keyword": "input_ids",
|
25 |
+
"sample_size": [
|
26 |
+
1,
|
27 |
+
32
|
28 |
+
],
|
29 |
+
"type": "long"
|
30 |
+
},
|
31 |
+
{
|
32 |
+
"keyword": "attention_mask",
|
33 |
+
"sample_size": [
|
34 |
+
1,
|
35 |
+
32
|
36 |
+
],
|
37 |
+
"type": "long"
|
38 |
+
}
|
39 |
+
],
|
40 |
+
"log_dir": "./logs/opt-6.7b-w8w8-unstructured50/",
|
41 |
+
"optimum_version": "1.13.2",
|
42 |
+
"save_onnx_model": false,
|
43 |
+
"transformers_version": "4.34.0"
|
44 |
+
}
|
openvino_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:78c59d0a9a9349be156fbe0cef16f776555f2a1293ffdf200be1579faa287468
|
3 |
+
size 7516461328
|
openvino_model.xml
ADDED
The diff for this file is too large to render.
See raw diff
|
|
original_graph.dot
ADDED
The diff for this file is too large to render.
See raw diff
|
|
ov_sparsity_stats.json
ADDED
@@ -0,0 +1,195 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"Constant_740364": 0.5002646446228027,
|
3 |
+
"Constant_740368": 0.5002545714378357,
|
4 |
+
"Constant_740372": 0.5001242160797119,
|
5 |
+
"Constant_740376": 0.5001725554466248,
|
6 |
+
"Constant_740380": 0.5000810772180557,
|
7 |
+
"Constant_740384": 0.50013767182827,
|
8 |
+
"Constant_740388": 0.5000916123390198,
|
9 |
+
"Constant_740392": 0.5001347064971924,
|
10 |
+
"Constant_740396": 0.5001572370529175,
|
11 |
+
"Constant_740400": 0.5001401305198669,
|
12 |
+
"Constant_740404": 0.5002889931201935,
|
13 |
+
"Constant_740408": 0.5001097619533539,
|
14 |
+
"Constant_740412": 0.5000396966934204,
|
15 |
+
"Constant_740416": 0.5000334978103638,
|
16 |
+
"Constant_740420": 0.500091016292572,
|
17 |
+
"Constant_740424": 0.5000038146972656,
|
18 |
+
"Constant_740428": 0.5001920312643051,
|
19 |
+
"Constant_740432": 0.5002889633178711,
|
20 |
+
"Constant_740436": 0.5001274347305298,
|
21 |
+
"Constant_740440": 0.5002991557121277,
|
22 |
+
"Constant_740444": 0.5000097155570984,
|
23 |
+
"Constant_740448": 0.5003159642219543,
|
24 |
+
"Constant_740452": 0.5002010613679886,
|
25 |
+
"Constant_740456": 0.5002553313970566,
|
26 |
+
"Constant_740460": 0.5001412630081177,
|
27 |
+
"Constant_740464": 0.5001715421676636,
|
28 |
+
"Constant_740468": 0.5000601410865784,
|
29 |
+
"Constant_740472": 0.5000429749488831,
|
30 |
+
"Constant_740476": 0.5002230852842331,
|
31 |
+
"Constant_740480": 0.5002836287021637,
|
32 |
+
"Constant_740484": 0.5001240372657776,
|
33 |
+
"Constant_740488": 0.5000519752502441,
|
34 |
+
"Constant_740492": 0.5000098347663879,
|
35 |
+
"Constant_740496": 0.5000232458114624,
|
36 |
+
"Constant_740500": 0.5000854283571243,
|
37 |
+
"Constant_740504": 0.5000220388174057,
|
38 |
+
"Constant_740508": 0.5000513792037964,
|
39 |
+
"Constant_740512": 0.5000736713409424,
|
40 |
+
"Constant_740516": 0.5001561045646667,
|
41 |
+
"Constant_740520": 0.5001461505889893,
|
42 |
+
"Constant_740524": 0.5000685006380081,
|
43 |
+
"Constant_740528": 0.5001946538686752,
|
44 |
+
"Constant_740532": 0.5001165866851807,
|
45 |
+
"Constant_740536": 0.5002720952033997,
|
46 |
+
"Constant_740540": 0.5002614259719849,
|
47 |
+
"Constant_740544": 0.5001602172851562,
|
48 |
+
"Constant_740548": 0.5002289712429047,
|
49 |
+
"Constant_740552": 0.5001602917909622,
|
50 |
+
"Constant_740556": 0.5000261068344116,
|
51 |
+
"Constant_740560": 0.500288188457489,
|
52 |
+
"Constant_740564": 0.5002854466438293,
|
53 |
+
"Constant_740568": 0.5000322461128235,
|
54 |
+
"Constant_740572": 0.5001158714294434,
|
55 |
+
"Constant_740576": 0.5001968443393707,
|
56 |
+
"Constant_740580": 0.5000325441360474,
|
57 |
+
"Constant_740584": 0.500220775604248,
|
58 |
+
"Constant_740588": 0.5002447366714478,
|
59 |
+
"Constant_740592": 0.5002323389053345,
|
60 |
+
"Constant_740596": 0.5001165121793747,
|
61 |
+
"Constant_740600": 0.5001854747533798,
|
62 |
+
"Constant_740604": 0.5000985264778137,
|
63 |
+
"Constant_740608": 0.5000840425491333,
|
64 |
+
"Constant_740612": 0.5001408457756042,
|
65 |
+
"Constant_740616": 0.5001469254493713,
|
66 |
+
"Constant_740620": 0.5002485513687134,
|
67 |
+
"Constant_740624": 0.5000925958156586,
|
68 |
+
"Constant_740628": 0.5001927614212036,
|
69 |
+
"Constant_740632": 0.5001641511917114,
|
70 |
+
"Constant_740636": 0.5002038478851318,
|
71 |
+
"Constant_740640": 0.5001091361045837,
|
72 |
+
"Constant_740644": 0.5001162588596344,
|
73 |
+
"Constant_740648": 0.5002866089344025,
|
74 |
+
"Constant_740652": 0.5001062154769897,
|
75 |
+
"Constant_740656": 0.5000807642936707,
|
76 |
+
"Constant_740660": 0.5003570318222046,
|
77 |
+
"Constant_740664": 0.5001456141471863,
|
78 |
+
"Constant_740668": 0.5002349317073822,
|
79 |
+
"Constant_740672": 0.500250443816185,
|
80 |
+
"Constant_740676": 0.5001181364059448,
|
81 |
+
"Constant_740680": 0.5002414584159851,
|
82 |
+
"Constant_740684": 0.5002016425132751,
|
83 |
+
"Constant_740688": 0.5001419186592102,
|
84 |
+
"Constant_740692": 0.5002192854881287,
|
85 |
+
"Constant_740696": 0.5002044141292572,
|
86 |
+
"Constant_740700": 0.5001293420791626,
|
87 |
+
"Constant_740704": 0.5000990629196167,
|
88 |
+
"Constant_740708": 0.5001548528671265,
|
89 |
+
"Constant_740712": 0.5000123381614685,
|
90 |
+
"Constant_740716": 0.5000330209732056,
|
91 |
+
"Constant_740720": 0.5000393390655518,
|
92 |
+
"Constant_740724": 0.5000688433647156,
|
93 |
+
"Constant_740728": 0.5001242160797119,
|
94 |
+
"Constant_740732": 0.5002279877662659,
|
95 |
+
"Constant_740736": 0.5000247359275818,
|
96 |
+
"Constant_740740": 0.500240683555603,
|
97 |
+
"Constant_740744": 0.5000388026237488,
|
98 |
+
"Constant_740748": 0.5001137852668762,
|
99 |
+
"Constant_740752": 0.5000342726707458,
|
100 |
+
"Constant_740756": 0.5001882314682007,
|
101 |
+
"Constant_740760": 0.5003361105918884,
|
102 |
+
"Constant_740764": 0.5002809017896652,
|
103 |
+
"Constant_740768": 0.5002386122941971,
|
104 |
+
"Constant_740772": 0.500178337097168,
|
105 |
+
"Constant_740776": 0.5001136660575867,
|
106 |
+
"Constant_740780": 0.5001077651977539,
|
107 |
+
"Constant_740784": 0.5003582239151001,
|
108 |
+
"Constant_740788": 0.5000877976417542,
|
109 |
+
"Constant_740792": 0.5002572685480118,
|
110 |
+
"Constant_740796": 0.5001864433288574,
|
111 |
+
"Constant_740800": 0.5001525282859802,
|
112 |
+
"Constant_740804": 0.5002452731132507,
|
113 |
+
"Constant_740808": 0.5003000497817993,
|
114 |
+
"Constant_740812": 0.500264897942543,
|
115 |
+
"Constant_740816": 0.5001865923404694,
|
116 |
+
"Constant_740820": 0.5002719759941101,
|
117 |
+
"Constant_740824": 0.5002580285072327,
|
118 |
+
"Constant_740828": 0.5001961588859558,
|
119 |
+
"Constant_740832": 0.5003326535224915,
|
120 |
+
"Constant_740836": 0.5000276118516922,
|
121 |
+
"Constant_740840": 0.500083327293396,
|
122 |
+
"Constant_740844": 0.5002673864364624,
|
123 |
+
"Constant_740848": 0.5000994801521301,
|
124 |
+
"Constant_740852": 0.5002595782279968,
|
125 |
+
"Constant_740856": 0.5001588463783264,
|
126 |
+
"Constant_740860": 0.5000343769788742,
|
127 |
+
"Constant_740864": 0.5002723783254623,
|
128 |
+
"Constant_740868": 0.5001829266548157,
|
129 |
+
"Constant_740872": 0.5002307295799255,
|
130 |
+
"Constant_740876": 0.5000202059745789,
|
131 |
+
"Constant_740880": 0.5001521110534668,
|
132 |
+
"Constant_740884": 0.5001227259635925,
|
133 |
+
"Constant_740888": 0.5002606809139252,
|
134 |
+
"Constant_740892": 0.5002914667129517,
|
135 |
+
"Constant_740896": 0.5001171827316284,
|
136 |
+
"Constant_740900": 0.5000696182250977,
|
137 |
+
"Constant_740904": 0.5002221465110779,
|
138 |
+
"Constant_740908": 0.5001393109560013,
|
139 |
+
"Constant_740912": 0.5002193599939346,
|
140 |
+
"Constant_740916": 0.5000544786453247,
|
141 |
+
"Constant_740920": 0.5001165866851807,
|
142 |
+
"Constant_740924": 0.5000243186950684,
|
143 |
+
"Constant_740928": 0.5000439286231995,
|
144 |
+
"Constant_740932": 0.5000893771648407,
|
145 |
+
"Constant_740936": 0.500079557299614,
|
146 |
+
"Constant_740940": 0.5002276301383972,
|
147 |
+
"Constant_740944": 0.500243067741394,
|
148 |
+
"Constant_740948": 0.5001072287559509,
|
149 |
+
"Constant_740952": 0.5000651478767395,
|
150 |
+
"Constant_740956": 0.5001714676618576,
|
151 |
+
"Constant_740960": 0.5000118166208267,
|
152 |
+
"Constant_740964": 0.5000373721122742,
|
153 |
+
"Constant_740968": 0.5002548694610596,
|
154 |
+
"Constant_740972": 0.5001973509788513,
|
155 |
+
"Constant_740976": 0.5001676082611084,
|
156 |
+
"Constant_740980": 0.5000311583280563,
|
157 |
+
"Constant_740984": 0.5000687539577484,
|
158 |
+
"Constant_740988": 0.5003244876861572,
|
159 |
+
"Constant_740992": 0.5001760125160217,
|
160 |
+
"Constant_740996": 0.5001277923583984,
|
161 |
+
"Constant_741000": 0.5002439618110657,
|
162 |
+
"Constant_741004": 0.5001821517944336,
|
163 |
+
"Constant_741008": 0.5001326352357864,
|
164 |
+
"Constant_741012": 0.5003024935722351,
|
165 |
+
"Constant_741016": 0.5000227093696594,
|
166 |
+
"Constant_741020": 0.5000532865524292,
|
167 |
+
"Constant_741024": 0.5000585913658142,
|
168 |
+
"Constant_741028": 0.500203400850296,
|
169 |
+
"Constant_741032": 0.500085711479187,
|
170 |
+
"Constant_741036": 0.5000766515731812,
|
171 |
+
"Constant_741040": 0.5001662373542786,
|
172 |
+
"Constant_741044": 0.5001732707023621,
|
173 |
+
"Constant_741048": 0.5001737475395203,
|
174 |
+
"Constant_741052": 0.5000820159912109,
|
175 |
+
"Constant_741056": 0.5001400411128998,
|
176 |
+
"Constant_741060": 0.5001081824302673,
|
177 |
+
"Constant_741064": 0.5001533627510071,
|
178 |
+
"Constant_741068": 0.5001651048660278,
|
179 |
+
"Constant_741072": 0.5000892877578735,
|
180 |
+
"Constant_741076": 0.5000907778739929,
|
181 |
+
"Constant_741080": 0.5000842660665512,
|
182 |
+
"Constant_741084": 0.5001571774482727,
|
183 |
+
"Constant_741088": 0.5000737309455872,
|
184 |
+
"Constant_741092": 0.5000970959663391,
|
185 |
+
"Constant_741096": 0.5000082850456238,
|
186 |
+
"Constant_741100": 0.5002415478229523,
|
187 |
+
"Constant_741104": 0.5001385509967804,
|
188 |
+
"Constant_741108": 0.5000050067901611,
|
189 |
+
"Constant_741112": 0.5001327395439148,
|
190 |
+
"Constant_741116": 0.500109076499939,
|
191 |
+
"Constant_741120": 0.500032901763916,
|
192 |
+
"Constant_741124": 0.5000126212835312,
|
193 |
+
"Constant_741128": 0.5000438243150711,
|
194 |
+
"Constant_741132": 0.013297121665949733
|
195 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "</s>",
|
3 |
+
"eos_token": "</s>",
|
4 |
+
"pad_token": "</s>",
|
5 |
+
"unk_token": "</s>"
|
6 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_prefix_space": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"1": {
|
6 |
+
"content": "<pad>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": true,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"2": {
|
14 |
+
"content": "</s>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": true,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
}
|
21 |
+
},
|
22 |
+
"additional_special_tokens": [],
|
23 |
+
"bos_token": "</s>",
|
24 |
+
"clean_up_tokenization_spaces": true,
|
25 |
+
"eos_token": "</s>",
|
26 |
+
"errors": "replace",
|
27 |
+
"model_max_length": 1000000000000000019884624838656,
|
28 |
+
"pad_token": "</s>",
|
29 |
+
"tokenizer_class": "GPT2Tokenizer",
|
30 |
+
"unk_token": "</s>"
|
31 |
+
}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|