Upload folder using huggingface_hub
Browse files- added_tokens.json +4 -0
- compressed_graph.dot +0 -0
- config.json +31 -0
- merges.txt +0 -0
- openvino_config.json +44 -0
- openvino_model.bin +3 -0
- openvino_model.xml +0 -0
- original_graph.dot +0 -0
- ov_sparsity_stats.json +149 -0
- special_tokens_map.json +6 -0
- tokenizer.json +0 -0
- tokenizer_config.json +31 -0
- vocab.json +0 -0
added_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"</s>": 2,
|
3 |
+
"<pad>": 1
|
4 |
+
}
|
compressed_graph.dot
ADDED
The diff for this file is too large to render.
See raw diff
|
|
config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "facebook/opt-350m",
|
3 |
+
"_remove_final_layer_norm": false,
|
4 |
+
"activation_dropout": 0.0,
|
5 |
+
"activation_function": "relu",
|
6 |
+
"architectures": [
|
7 |
+
"OPTForCausalLM"
|
8 |
+
],
|
9 |
+
"attention_dropout": 0.0,
|
10 |
+
"bos_token_id": 2,
|
11 |
+
"do_layer_norm_before": false,
|
12 |
+
"dropout": 0.1,
|
13 |
+
"enable_bias": true,
|
14 |
+
"eos_token_id": 2,
|
15 |
+
"ffn_dim": 4096,
|
16 |
+
"hidden_size": 1024,
|
17 |
+
"init_std": 0.02,
|
18 |
+
"layer_norm_elementwise_affine": true,
|
19 |
+
"layerdrop": 0.0,
|
20 |
+
"max_position_embeddings": 2048,
|
21 |
+
"model_type": "opt",
|
22 |
+
"num_attention_heads": 16,
|
23 |
+
"num_hidden_layers": 24,
|
24 |
+
"pad_token_id": 1,
|
25 |
+
"prefix": "</s>",
|
26 |
+
"torch_dtype": "float16",
|
27 |
+
"transformers_version": "4.34.0",
|
28 |
+
"use_cache": true,
|
29 |
+
"vocab_size": 50272,
|
30 |
+
"word_embed_proj_dim": 512
|
31 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
openvino_config.json
ADDED
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"compression": {
|
3 |
+
"algorithm": "quantization",
|
4 |
+
"export_to_onnx_standard_ops": false,
|
5 |
+
"ignored_scopes": [
|
6 |
+
"{re}.*Embedding.*",
|
7 |
+
"{re}.*add___.*",
|
8 |
+
"{re}.*layer_norm_.*"
|
9 |
+
],
|
10 |
+
"initializer": {
|
11 |
+
"batchnorm_adaptation": {
|
12 |
+
"num_bn_adaptation_samples": 0
|
13 |
+
},
|
14 |
+
"range": {
|
15 |
+
"num_init_samples": 16,
|
16 |
+
"type": "min_max"
|
17 |
+
}
|
18 |
+
},
|
19 |
+
"overflow_fix": "disable",
|
20 |
+
"preset": "performance"
|
21 |
+
},
|
22 |
+
"input_info": [
|
23 |
+
{
|
24 |
+
"keyword": "input_ids",
|
25 |
+
"sample_size": [
|
26 |
+
1,
|
27 |
+
32
|
28 |
+
],
|
29 |
+
"type": "long"
|
30 |
+
},
|
31 |
+
{
|
32 |
+
"keyword": "attention_mask",
|
33 |
+
"sample_size": [
|
34 |
+
1,
|
35 |
+
32
|
36 |
+
],
|
37 |
+
"type": "long"
|
38 |
+
}
|
39 |
+
],
|
40 |
+
"log_dir": "./logs/opt-350m-w8w8-unstructured50/",
|
41 |
+
"optimum_version": "1.13.2",
|
42 |
+
"save_onnx_model": false,
|
43 |
+
"transformers_version": "4.34.0"
|
44 |
+
}
|
openvino_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dfcff4277cc24fcb5853341ca033bdf3b43f14b0a6e7c796e4eaea81ee3c23d3
|
3 |
+
size 442306456
|
openvino_model.xml
ADDED
The diff for this file is too large to render.
See raw diff
|
|
original_graph.dot
ADDED
The diff for this file is too large to render.
See raw diff
|
|
ov_sparsity_stats.json
ADDED
@@ -0,0 +1,149 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"Constant_201525": 0.0069179534912109375,
|
3 |
+
"Constant_201529": 0.50006103515625,
|
4 |
+
"Constant_201533": 0.5007028579711914,
|
5 |
+
"Constant_201537": 0.5001296997070312,
|
6 |
+
"Constant_201541": 0.5001859664916992,
|
7 |
+
"Constant_201545": 0.50016188621521,
|
8 |
+
"Constant_201549": 0.500258207321167,
|
9 |
+
"Constant_201553": 0.5001916885375977,
|
10 |
+
"Constant_201557": 0.5002012252807617,
|
11 |
+
"Constant_201561": 0.5000591278076172,
|
12 |
+
"Constant_201565": 0.5001449584960938,
|
13 |
+
"Constant_201569": 0.5001871585845947,
|
14 |
+
"Constant_201573": 0.5000674724578857,
|
15 |
+
"Constant_201577": 0.5001707077026367,
|
16 |
+
"Constant_201581": 0.5001487731933594,
|
17 |
+
"Constant_201585": 0.5000381469726562,
|
18 |
+
"Constant_201589": 0.5000143051147461,
|
19 |
+
"Constant_201593": 0.5002789497375488,
|
20 |
+
"Constant_201597": 0.5002901554107666,
|
21 |
+
"Constant_201601": 0.5001840591430664,
|
22 |
+
"Constant_201605": 0.5001916885375977,
|
23 |
+
"Constant_201609": 0.5003271102905273,
|
24 |
+
"Constant_201613": 0.5000181198120117,
|
25 |
+
"Constant_201617": 0.5001442432403564,
|
26 |
+
"Constant_201621": 0.5001802444458008,
|
27 |
+
"Constant_201625": 0.5000400543212891,
|
28 |
+
"Constant_201629": 0.5001459121704102,
|
29 |
+
"Constant_201633": 0.5001134872436523,
|
30 |
+
"Constant_201637": 0.5000715255737305,
|
31 |
+
"Constant_201641": 0.500129222869873,
|
32 |
+
"Constant_201645": 0.5002434253692627,
|
33 |
+
"Constant_201649": 0.5000200271606445,
|
34 |
+
"Constant_201653": 0.500244140625,
|
35 |
+
"Constant_201657": 0.5000839233398438,
|
36 |
+
"Constant_201661": 0.5002164840698242,
|
37 |
+
"Constant_201665": 0.5000641345977783,
|
38 |
+
"Constant_201669": 0.5002322196960449,
|
39 |
+
"Constant_201673": 0.5000286102294922,
|
40 |
+
"Constant_201677": 0.5002031326293945,
|
41 |
+
"Constant_201681": 0.5005426406860352,
|
42 |
+
"Constant_201685": 0.5001363754272461,
|
43 |
+
"Constant_201689": 0.5001888275146484,
|
44 |
+
"Constant_201693": 0.500004768371582,
|
45 |
+
"Constant_201697": 0.5000076293945312,
|
46 |
+
"Constant_201701": 0.5001287460327148,
|
47 |
+
"Constant_201705": 0.5004806518554688,
|
48 |
+
"Constant_201709": 0.5001335144042969,
|
49 |
+
"Constant_201713": 0.5000123977661133,
|
50 |
+
"Constant_201717": 0.5001475811004639,
|
51 |
+
"Constant_201721": 0.5000696182250977,
|
52 |
+
"Constant_201725": 0.5002098083496094,
|
53 |
+
"Constant_201729": 0.5004663467407227,
|
54 |
+
"Constant_201733": 0.5002079010009766,
|
55 |
+
"Constant_201737": 0.5002341270446777,
|
56 |
+
"Constant_201741": 0.5000836849212646,
|
57 |
+
"Constant_201745": 0.5001115798950195,
|
58 |
+
"Constant_201749": 0.5000095367431641,
|
59 |
+
"Constant_201753": 0.5002813339233398,
|
60 |
+
"Constant_201757": 0.5000953674316406,
|
61 |
+
"Constant_201761": 0.5000920295715332,
|
62 |
+
"Constant_201765": 0.5000362396240234,
|
63 |
+
"Constant_201769": 0.500056266784668,
|
64 |
+
"Constant_201773": 0.5001335144042969,
|
65 |
+
"Constant_201777": 0.5001678466796875,
|
66 |
+
"Constant_201781": 0.5002002716064453,
|
67 |
+
"Constant_201785": 0.5001099109649658,
|
68 |
+
"Constant_201789": 0.5000040531158447,
|
69 |
+
"Constant_201793": 0.5000696182250977,
|
70 |
+
"Constant_201797": 0.5001897811889648,
|
71 |
+
"Constant_201801": 0.500213623046875,
|
72 |
+
"Constant_201805": 0.5000772476196289,
|
73 |
+
"Constant_201809": 0.5000016689300537,
|
74 |
+
"Constant_201813": 0.5001311302185059,
|
75 |
+
"Constant_201817": 0.5000152587890625,
|
76 |
+
"Constant_201821": 0.5001249313354492,
|
77 |
+
"Constant_201825": 0.5000734329223633,
|
78 |
+
"Constant_201829": 0.5003366470336914,
|
79 |
+
"Constant_201833": 0.5001416206359863,
|
80 |
+
"Constant_201837": 0.5001189708709717,
|
81 |
+
"Constant_201841": 0.5002670288085938,
|
82 |
+
"Constant_201845": 0.5002050399780273,
|
83 |
+
"Constant_201849": 0.5002527236938477,
|
84 |
+
"Constant_201853": 0.5001449584960938,
|
85 |
+
"Constant_201857": 0.5001153945922852,
|
86 |
+
"Constant_201861": 0.5001699924468994,
|
87 |
+
"Constant_201865": 0.5001010894775391,
|
88 |
+
"Constant_201869": 0.5001974105834961,
|
89 |
+
"Constant_201873": 0.500030517578125,
|
90 |
+
"Constant_201877": 0.5001068115234375,
|
91 |
+
"Constant_201881": 0.5000848770141602,
|
92 |
+
"Constant_201885": 0.5000336170196533,
|
93 |
+
"Constant_201889": 0.5002536773681641,
|
94 |
+
"Constant_201893": 0.5002059936523438,
|
95 |
+
"Constant_201897": 0.5001115798950195,
|
96 |
+
"Constant_201901": 0.5000457763671875,
|
97 |
+
"Constant_201905": 0.5002875328063965,
|
98 |
+
"Constant_201909": 0.500126838684082,
|
99 |
+
"Constant_201913": 0.5001535415649414,
|
100 |
+
"Constant_201917": 0.5001840591430664,
|
101 |
+
"Constant_201921": 0.500187873840332,
|
102 |
+
"Constant_201925": 0.5001668930053711,
|
103 |
+
"Constant_201929": 0.500324010848999,
|
104 |
+
"Constant_201933": 0.5001082420349121,
|
105 |
+
"Constant_201937": 0.5002193450927734,
|
106 |
+
"Constant_201941": 0.5001583099365234,
|
107 |
+
"Constant_201945": 0.5000391006469727,
|
108 |
+
"Constant_201949": 0.5000143051147461,
|
109 |
+
"Constant_201953": 0.5001306533813477,
|
110 |
+
"Constant_201957": 0.5001113414764404,
|
111 |
+
"Constant_201961": 0.5001735687255859,
|
112 |
+
"Constant_201965": 0.5001688003540039,
|
113 |
+
"Constant_201969": 0.5000753402709961,
|
114 |
+
"Constant_201973": 0.5002660751342773,
|
115 |
+
"Constant_201977": 0.5000832080841064,
|
116 |
+
"Constant_201981": 0.5001823902130127,
|
117 |
+
"Constant_201985": 0.5002040863037109,
|
118 |
+
"Constant_201989": 0.5001678466796875,
|
119 |
+
"Constant_201993": 0.5002288818359375,
|
120 |
+
"Constant_201997": 0.5000133514404297,
|
121 |
+
"Constant_202001": 0.5002453327178955,
|
122 |
+
"Constant_202005": 0.5000607967376709,
|
123 |
+
"Constant_202009": 0.5000705718994141,
|
124 |
+
"Constant_202013": 0.5000247955322266,
|
125 |
+
"Constant_202017": 0.5001020431518555,
|
126 |
+
"Constant_202021": 0.5003366470336914,
|
127 |
+
"Constant_202025": 0.5000171661376953,
|
128 |
+
"Constant_202029": 0.5000460147857666,
|
129 |
+
"Constant_202033": 0.5002079010009766,
|
130 |
+
"Constant_202037": 0.5001144409179688,
|
131 |
+
"Constant_202041": 0.5002431869506836,
|
132 |
+
"Constant_202045": 0.5000209808349609,
|
133 |
+
"Constant_202049": 0.5001280307769775,
|
134 |
+
"Constant_202053": 0.5001664161682129,
|
135 |
+
"Constant_202057": 0.5001163482666016,
|
136 |
+
"Constant_202061": 0.5002288818359375,
|
137 |
+
"Constant_202065": 0.5002784729003906,
|
138 |
+
"Constant_202069": 0.5000295639038086,
|
139 |
+
"Constant_202073": 0.500084400177002,
|
140 |
+
"Constant_202077": 0.5001175403594971,
|
141 |
+
"Constant_202081": 0.5000619888305664,
|
142 |
+
"Constant_202085": 0.5002574920654297,
|
143 |
+
"Constant_202089": 0.5000839233398438,
|
144 |
+
"Constant_202093": 0.5000972747802734,
|
145 |
+
"Constant_202097": 0.5002822875976562,
|
146 |
+
"Constant_202101": 0.5000905990600586,
|
147 |
+
"Constant_202105": 0.0028743743896484375,
|
148 |
+
"Constant_202109": 0.008723831419577499
|
149 |
+
}
|
special_tokens_map.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": "</s>",
|
3 |
+
"eos_token": "</s>",
|
4 |
+
"pad_token": "</s>",
|
5 |
+
"unk_token": "</s>"
|
6 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_prefix_space": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"1": {
|
6 |
+
"content": "<pad>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": true,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"2": {
|
14 |
+
"content": "</s>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": true,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
}
|
21 |
+
},
|
22 |
+
"additional_special_tokens": [],
|
23 |
+
"bos_token": "</s>",
|
24 |
+
"clean_up_tokenization_spaces": true,
|
25 |
+
"eos_token": "</s>",
|
26 |
+
"errors": "replace",
|
27 |
+
"model_max_length": 1000000000000000019884624838656,
|
28 |
+
"pad_token": "</s>",
|
29 |
+
"tokenizer_class": "GPT2Tokenizer",
|
30 |
+
"unk_token": "</s>"
|
31 |
+
}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|