yujiepan commited on
Commit
850e970
1 Parent(s): f0b0b2a

Upload folder using huggingface_hub

Browse files
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "</s>": 2,
3
+ "<pad>": 1
4
+ }
compressed_graph.dot ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/opt-350m",
3
+ "_remove_final_layer_norm": false,
4
+ "activation_dropout": 0.0,
5
+ "activation_function": "relu",
6
+ "architectures": [
7
+ "OPTForCausalLM"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "bos_token_id": 2,
11
+ "do_layer_norm_before": false,
12
+ "dropout": 0.1,
13
+ "enable_bias": true,
14
+ "eos_token_id": 2,
15
+ "ffn_dim": 4096,
16
+ "hidden_size": 1024,
17
+ "init_std": 0.02,
18
+ "layer_norm_elementwise_affine": true,
19
+ "layerdrop": 0.0,
20
+ "max_position_embeddings": 2048,
21
+ "model_type": "opt",
22
+ "num_attention_heads": 16,
23
+ "num_hidden_layers": 24,
24
+ "pad_token_id": 1,
25
+ "prefix": "</s>",
26
+ "torch_dtype": "float16",
27
+ "transformers_version": "4.34.0",
28
+ "use_cache": true,
29
+ "vocab_size": 50272,
30
+ "word_embed_proj_dim": 512
31
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
openvino_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "compression": {
3
+ "algorithm": "quantization",
4
+ "export_to_onnx_standard_ops": false,
5
+ "ignored_scopes": [
6
+ "{re}.*Embedding.*",
7
+ "{re}.*add___.*",
8
+ "{re}.*layer_norm_.*"
9
+ ],
10
+ "initializer": {
11
+ "batchnorm_adaptation": {
12
+ "num_bn_adaptation_samples": 0
13
+ },
14
+ "range": {
15
+ "num_init_samples": 16,
16
+ "type": "min_max"
17
+ }
18
+ },
19
+ "overflow_fix": "disable",
20
+ "preset": "performance"
21
+ },
22
+ "input_info": [
23
+ {
24
+ "keyword": "input_ids",
25
+ "sample_size": [
26
+ 1,
27
+ 32
28
+ ],
29
+ "type": "long"
30
+ },
31
+ {
32
+ "keyword": "attention_mask",
33
+ "sample_size": [
34
+ 1,
35
+ 32
36
+ ],
37
+ "type": "long"
38
+ }
39
+ ],
40
+ "log_dir": "./logs/opt-350m-w8w8-unstructured50/",
41
+ "optimum_version": "1.13.2",
42
+ "save_onnx_model": false,
43
+ "transformers_version": "4.34.0"
44
+ }
openvino_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfcff4277cc24fcb5853341ca033bdf3b43f14b0a6e7c796e4eaea81ee3c23d3
3
+ size 442306456
openvino_model.xml ADDED
The diff for this file is too large to render. See raw diff
 
original_graph.dot ADDED
The diff for this file is too large to render. See raw diff
 
ov_sparsity_stats.json ADDED
@@ -0,0 +1,149 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "Constant_201525": 0.0069179534912109375,
3
+ "Constant_201529": 0.50006103515625,
4
+ "Constant_201533": 0.5007028579711914,
5
+ "Constant_201537": 0.5001296997070312,
6
+ "Constant_201541": 0.5001859664916992,
7
+ "Constant_201545": 0.50016188621521,
8
+ "Constant_201549": 0.500258207321167,
9
+ "Constant_201553": 0.5001916885375977,
10
+ "Constant_201557": 0.5002012252807617,
11
+ "Constant_201561": 0.5000591278076172,
12
+ "Constant_201565": 0.5001449584960938,
13
+ "Constant_201569": 0.5001871585845947,
14
+ "Constant_201573": 0.5000674724578857,
15
+ "Constant_201577": 0.5001707077026367,
16
+ "Constant_201581": 0.5001487731933594,
17
+ "Constant_201585": 0.5000381469726562,
18
+ "Constant_201589": 0.5000143051147461,
19
+ "Constant_201593": 0.5002789497375488,
20
+ "Constant_201597": 0.5002901554107666,
21
+ "Constant_201601": 0.5001840591430664,
22
+ "Constant_201605": 0.5001916885375977,
23
+ "Constant_201609": 0.5003271102905273,
24
+ "Constant_201613": 0.5000181198120117,
25
+ "Constant_201617": 0.5001442432403564,
26
+ "Constant_201621": 0.5001802444458008,
27
+ "Constant_201625": 0.5000400543212891,
28
+ "Constant_201629": 0.5001459121704102,
29
+ "Constant_201633": 0.5001134872436523,
30
+ "Constant_201637": 0.5000715255737305,
31
+ "Constant_201641": 0.500129222869873,
32
+ "Constant_201645": 0.5002434253692627,
33
+ "Constant_201649": 0.5000200271606445,
34
+ "Constant_201653": 0.500244140625,
35
+ "Constant_201657": 0.5000839233398438,
36
+ "Constant_201661": 0.5002164840698242,
37
+ "Constant_201665": 0.5000641345977783,
38
+ "Constant_201669": 0.5002322196960449,
39
+ "Constant_201673": 0.5000286102294922,
40
+ "Constant_201677": 0.5002031326293945,
41
+ "Constant_201681": 0.5005426406860352,
42
+ "Constant_201685": 0.5001363754272461,
43
+ "Constant_201689": 0.5001888275146484,
44
+ "Constant_201693": 0.500004768371582,
45
+ "Constant_201697": 0.5000076293945312,
46
+ "Constant_201701": 0.5001287460327148,
47
+ "Constant_201705": 0.5004806518554688,
48
+ "Constant_201709": 0.5001335144042969,
49
+ "Constant_201713": 0.5000123977661133,
50
+ "Constant_201717": 0.5001475811004639,
51
+ "Constant_201721": 0.5000696182250977,
52
+ "Constant_201725": 0.5002098083496094,
53
+ "Constant_201729": 0.5004663467407227,
54
+ "Constant_201733": 0.5002079010009766,
55
+ "Constant_201737": 0.5002341270446777,
56
+ "Constant_201741": 0.5000836849212646,
57
+ "Constant_201745": 0.5001115798950195,
58
+ "Constant_201749": 0.5000095367431641,
59
+ "Constant_201753": 0.5002813339233398,
60
+ "Constant_201757": 0.5000953674316406,
61
+ "Constant_201761": 0.5000920295715332,
62
+ "Constant_201765": 0.5000362396240234,
63
+ "Constant_201769": 0.500056266784668,
64
+ "Constant_201773": 0.5001335144042969,
65
+ "Constant_201777": 0.5001678466796875,
66
+ "Constant_201781": 0.5002002716064453,
67
+ "Constant_201785": 0.5001099109649658,
68
+ "Constant_201789": 0.5000040531158447,
69
+ "Constant_201793": 0.5000696182250977,
70
+ "Constant_201797": 0.5001897811889648,
71
+ "Constant_201801": 0.500213623046875,
72
+ "Constant_201805": 0.5000772476196289,
73
+ "Constant_201809": 0.5000016689300537,
74
+ "Constant_201813": 0.5001311302185059,
75
+ "Constant_201817": 0.5000152587890625,
76
+ "Constant_201821": 0.5001249313354492,
77
+ "Constant_201825": 0.5000734329223633,
78
+ "Constant_201829": 0.5003366470336914,
79
+ "Constant_201833": 0.5001416206359863,
80
+ "Constant_201837": 0.5001189708709717,
81
+ "Constant_201841": 0.5002670288085938,
82
+ "Constant_201845": 0.5002050399780273,
83
+ "Constant_201849": 0.5002527236938477,
84
+ "Constant_201853": 0.5001449584960938,
85
+ "Constant_201857": 0.5001153945922852,
86
+ "Constant_201861": 0.5001699924468994,
87
+ "Constant_201865": 0.5001010894775391,
88
+ "Constant_201869": 0.5001974105834961,
89
+ "Constant_201873": 0.500030517578125,
90
+ "Constant_201877": 0.5001068115234375,
91
+ "Constant_201881": 0.5000848770141602,
92
+ "Constant_201885": 0.5000336170196533,
93
+ "Constant_201889": 0.5002536773681641,
94
+ "Constant_201893": 0.5002059936523438,
95
+ "Constant_201897": 0.5001115798950195,
96
+ "Constant_201901": 0.5000457763671875,
97
+ "Constant_201905": 0.5002875328063965,
98
+ "Constant_201909": 0.500126838684082,
99
+ "Constant_201913": 0.5001535415649414,
100
+ "Constant_201917": 0.5001840591430664,
101
+ "Constant_201921": 0.500187873840332,
102
+ "Constant_201925": 0.5001668930053711,
103
+ "Constant_201929": 0.500324010848999,
104
+ "Constant_201933": 0.5001082420349121,
105
+ "Constant_201937": 0.5002193450927734,
106
+ "Constant_201941": 0.5001583099365234,
107
+ "Constant_201945": 0.5000391006469727,
108
+ "Constant_201949": 0.5000143051147461,
109
+ "Constant_201953": 0.5001306533813477,
110
+ "Constant_201957": 0.5001113414764404,
111
+ "Constant_201961": 0.5001735687255859,
112
+ "Constant_201965": 0.5001688003540039,
113
+ "Constant_201969": 0.5000753402709961,
114
+ "Constant_201973": 0.5002660751342773,
115
+ "Constant_201977": 0.5000832080841064,
116
+ "Constant_201981": 0.5001823902130127,
117
+ "Constant_201985": 0.5002040863037109,
118
+ "Constant_201989": 0.5001678466796875,
119
+ "Constant_201993": 0.5002288818359375,
120
+ "Constant_201997": 0.5000133514404297,
121
+ "Constant_202001": 0.5002453327178955,
122
+ "Constant_202005": 0.5000607967376709,
123
+ "Constant_202009": 0.5000705718994141,
124
+ "Constant_202013": 0.5000247955322266,
125
+ "Constant_202017": 0.5001020431518555,
126
+ "Constant_202021": 0.5003366470336914,
127
+ "Constant_202025": 0.5000171661376953,
128
+ "Constant_202029": 0.5000460147857666,
129
+ "Constant_202033": 0.5002079010009766,
130
+ "Constant_202037": 0.5001144409179688,
131
+ "Constant_202041": 0.5002431869506836,
132
+ "Constant_202045": 0.5000209808349609,
133
+ "Constant_202049": 0.5001280307769775,
134
+ "Constant_202053": 0.5001664161682129,
135
+ "Constant_202057": 0.5001163482666016,
136
+ "Constant_202061": 0.5002288818359375,
137
+ "Constant_202065": 0.5002784729003906,
138
+ "Constant_202069": 0.5000295639038086,
139
+ "Constant_202073": 0.500084400177002,
140
+ "Constant_202077": 0.5001175403594971,
141
+ "Constant_202081": 0.5000619888305664,
142
+ "Constant_202085": 0.5002574920654297,
143
+ "Constant_202089": 0.5000839233398438,
144
+ "Constant_202093": 0.5000972747802734,
145
+ "Constant_202097": 0.5002822875976562,
146
+ "Constant_202101": 0.5000905990600586,
147
+ "Constant_202105": 0.0028743743896484375,
148
+ "Constant_202109": 0.008723831419577499
149
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "</s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "</s>",
5
+ "unk_token": "</s>"
6
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "1": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "2": {
14
+ "content": "</s>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ }
21
+ },
22
+ "additional_special_tokens": [],
23
+ "bos_token": "</s>",
24
+ "clean_up_tokenization_spaces": true,
25
+ "eos_token": "</s>",
26
+ "errors": "replace",
27
+ "model_max_length": 1000000000000000019884624838656,
28
+ "pad_token": "</s>",
29
+ "tokenizer_class": "GPT2Tokenizer",
30
+ "unk_token": "</s>"
31
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff