Felix Marty commited on
Commit
2d16027
1 Parent(s): d7bbae3
Files changed (4) hide show
  1. README.md +14 -0
  2. eval_results.json +3 -0
  3. model.onnx +3 -0
  4. ort_config.json +190 -0
README.md CHANGED
@@ -1,3 +1,17 @@
1
  ---
2
  license: apache-2.0
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - sst2
5
+ - glue
6
  ---
7
+
8
+ This model is a fork of https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english , quantized using static Post-Training Quantization (PTQ) with ONNX Runtime and 🤗 Optimum library.
9
+
10
+ It achieves 0.901 on the validation set.
11
+
12
+ To load this model:
13
+
14
+ ```python
15
+ from optimum.onnxruntime import ORTModelForSequenceClassification
16
+ model = ORTModelForSequenceClassification.from_pretrained("fxmarty/distilbert-base-uncased-finetuned-sst-2-english-int8-static")
17
+ ```
eval_results.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "accuracy": 0.8944954128440367
3
+ }
model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5c9b729a25e79ed70c1f885f6a0d04c5499370267159ee7b13ee4160d11a619
3
+ size 140444595
ort_config.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "opset": null,
3
+ "optimization": {},
4
+ "optimum_version": "1.4.1.dev0",
5
+ "quantization": {
6
+ "activations_dtype": "QInt8",
7
+ "activations_symmetric": false,
8
+ "format": "QDQ",
9
+ "is_static": true,
10
+ "mode": "QLinearOps",
11
+ "nodes_to_exclude": [
12
+ "Mul_503",
13
+ "Div_238",
14
+ "Mul_257",
15
+ "Add_316",
16
+ "ReduceMean_252",
17
+ "Add_575",
18
+ "Sqrt_91",
19
+ "Add_476",
20
+ "Mul_175",
21
+ "Pow_497",
22
+ "Mul_561",
23
+ "Mul_151",
24
+ "Pow_333",
25
+ "ReduceMean_416",
26
+ "Sub_85",
27
+ "Add_340",
28
+ "Erf_239",
29
+ "Add_387",
30
+ "Div_256",
31
+ "Add_487",
32
+ "Add_230",
33
+ "Pow_555",
34
+ "Add_411",
35
+ "Sqrt_231",
36
+ "Div_338",
37
+ "Add_312",
38
+ "Pow_169",
39
+ "Sub_471",
40
+ "Erf_321",
41
+ "Add_405",
42
+ "Add_159",
43
+ "Mul_488",
44
+ "Sqrt_313",
45
+ "Div_484",
46
+ "ReduceMean_170",
47
+ "ReduceMean_248",
48
+ "Mul_242",
49
+ "Add_247",
50
+ "Mul_572",
51
+ "Div_402",
52
+ "Add_493",
53
+ "ReduceMean_498",
54
+ "ReduceMean_494",
55
+ "ReduceMean_224",
56
+ "Div_150",
57
+ "Pow_145",
58
+ "ReduceMean_412",
59
+ "Sqrt_395",
60
+ "Mul_93",
61
+ "Add_223",
62
+ "Sqrt_149",
63
+ "Div_566",
64
+ "Sqrt_559",
65
+ "Add_394",
66
+ "Sqrt_477",
67
+ "Mul_160",
68
+ "Pow_251",
69
+ "Add_562",
70
+ "Add_569",
71
+ "Sub_143",
72
+ "Div_396",
73
+ "Mul_233",
74
+ "Mul_324",
75
+ "Erf_403",
76
+ "ReduceMean_580",
77
+ "Sub_331",
78
+ "Pow_87",
79
+ "Erf_485",
80
+ "Add_234",
81
+ "Sub_389",
82
+ "ReduceMean_330",
83
+ "Div_92",
84
+ "Sqrt_255",
85
+ "ReduceMean_228",
86
+ "Mul_315",
87
+ "Sub_167",
88
+ "Add_336",
89
+ "Add_172",
90
+ "Pow_473",
91
+ "ReduceMean_310",
92
+ "Mul_162",
93
+ "Mul_339",
94
+ "Add_551",
95
+ "Add_422",
96
+ "Sub_553",
97
+ "Mul_479",
98
+ "ReduceMean_306",
99
+ "Div_502",
100
+ "Pow_579",
101
+ "Add_582",
102
+ "Div_420",
103
+ "ReduceMean_166",
104
+ "Add_418",
105
+ "ReduceMean_142",
106
+ "Div_478",
107
+ "ReduceMean_470",
108
+ "Add_241",
109
+ "Mul_326",
110
+ "Mul_397",
111
+ "ReduceMean_88",
112
+ "Div_174",
113
+ "ReduceMean_556",
114
+ "Sub_577",
115
+ "Add_258",
116
+ "Sqrt_419",
117
+ "Div_560",
118
+ "ReduceMean_576",
119
+ "ReduceMean_388",
120
+ "ReduceMean_146",
121
+ "Sub_249",
122
+ "Pow_309",
123
+ "Sub_495",
124
+ "Sqrt_501",
125
+ "Add_152",
126
+ "Add_504",
127
+ "Sub_413",
128
+ "ReduceMean_552",
129
+ "Mul_408",
130
+ "Add_148",
131
+ "Sqrt_173",
132
+ "Sub_225",
133
+ "Add_254",
134
+ "Add_500",
135
+ "Add_398",
136
+ "Add_558",
137
+ "Sqrt_583",
138
+ "Add_83",
139
+ "Erf_567",
140
+ "Sqrt_337",
141
+ "Mul_490",
142
+ "Sub_307",
143
+ "Add_329",
144
+ "Div_314",
145
+ "Add_165",
146
+ "ReduceMean_392",
147
+ "Add_480",
148
+ "Pow_227",
149
+ "ReduceMean_334",
150
+ "ReduceMean_474",
151
+ "Add_141",
152
+ "Mul_406",
153
+ "Add_323",
154
+ "Add_469",
155
+ "Mul_585",
156
+ "Div_320",
157
+ "Div_232",
158
+ "Mul_421",
159
+ "Add_176",
160
+ "Mul_244",
161
+ "Add_586",
162
+ "Mul_570",
163
+ "Add_94",
164
+ "Pow_391",
165
+ "ReduceMean_84",
166
+ "Pow_415",
167
+ "Add_305",
168
+ "Div_584",
169
+ "Erf_157",
170
+ "Div_156",
171
+ "Add_90"
172
+ ],
173
+ "nodes_to_quantize": [],
174
+ "operators_to_quantize": [
175
+ "MatMul",
176
+ "Add"
177
+ ],
178
+ "per_channel": false,
179
+ "qdq_add_pair_to_weight": false,
180
+ "qdq_dedicated_pair": false,
181
+ "qdq_op_type_per_channel_support_to_axis": {
182
+ "MatMul": 1
183
+ },
184
+ "reduce_range": false,
185
+ "weights_dtype": "QInt8",
186
+ "weights_symmetric": true
187
+ },
188
+ "transformers_version": "4.23.0.dev0",
189
+ "use_external_data_format": false
190
+ }