Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json
CHANGED
@@ -6,19 +6,17 @@
|
|
6 |
"version": "2.3.0+cu121",
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "multiple-choice",
|
9 |
-
"model": "FacebookAI/roberta-base",
|
10 |
"library": "transformers",
|
|
|
|
|
11 |
"device": "cuda",
|
12 |
"device_ids": "0",
|
13 |
"seed": 42,
|
14 |
"inter_op_num_threads": null,
|
15 |
"intra_op_num_threads": null,
|
16 |
-
"
|
17 |
-
|
18 |
-
|
19 |
-
"local_files_only": false,
|
20 |
-
"trust_remote_code": false
|
21 |
-
},
|
22 |
"no_weights": true,
|
23 |
"device_map": null,
|
24 |
"torch_dtype": null,
|
@@ -104,7 +102,7 @@
|
|
104 |
"forward": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
-
"max_ram": 909.
|
108 |
"max_global_vram": 1195.900928,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 555.74528,
|
@@ -112,163 +110,160 @@
|
|
112 |
},
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
-
"count":
|
116 |
-
"total":
|
117 |
-
"mean": 0.
|
118 |
-
"stdev": 0.
|
119 |
-
"p50": 0.
|
120 |
-
"p90": 0.
|
121 |
-
"p95": 0.
|
122 |
-
"p99": 0.
|
123 |
"values": [
|
124 |
-
0.
|
125 |
-
0.
|
126 |
-
0.
|
127 |
-
0.
|
128 |
-
0.
|
129 |
-
0.
|
130 |
-
0.
|
131 |
-
0.
|
132 |
-
0.
|
133 |
-
0.
|
134 |
-
0.
|
135 |
-
0.
|
136 |
-
0.0078919677734375,
|
137 |
-
0.007888895988464355,
|
138 |
-
0.008018943786621094,
|
139 |
-
0.007961599826812745,
|
140 |
-
0.008143872261047362,
|
141 |
-
0.007972864151000977,
|
142 |
-
0.007878623962402343,
|
143 |
-
0.00786636781692505,
|
144 |
-
0.007832575798034667,
|
145 |
-
0.007867392063140868,
|
146 |
-
0.007902207851409913,
|
147 |
-
0.007871488094329833,
|
148 |
-
0.007935999870300293,
|
149 |
-
0.007916543960571289,
|
150 |
-
0.007854080200195313,
|
151 |
-
0.007862271785736084,
|
152 |
-
0.007865344047546387,
|
153 |
-
0.00787660789489746,
|
154 |
-
0.007796703815460205,
|
155 |
-
0.00773529577255249,
|
156 |
-
0.0077761921882629394,
|
157 |
-
0.007821311950683594,
|
158 |
-
0.007766016006469726,
|
159 |
-
0.007824384212493896,
|
160 |
-
0.007837696075439453,
|
161 |
-
0.007839744091033935,
|
162 |
0.007942143917083741,
|
163 |
-
0.
|
164 |
-
0.
|
165 |
-
0.
|
166 |
-
0.
|
167 |
-
0.
|
168 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
169 |
0.00795136022567749,
|
170 |
-
0.
|
171 |
-
0.
|
172 |
-
0.
|
173 |
-
0.
|
174 |
-
0.
|
175 |
-
0.
|
176 |
-
0.
|
177 |
-
0.
|
178 |
-
0.
|
179 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
180 |
0.007832575798034667,
|
181 |
-
0.
|
182 |
-
0.
|
183 |
-
0.
|
184 |
-
0.
|
185 |
-
0.
|
186 |
-
0.
|
187 |
-
0.
|
188 |
-
0.
|
189 |
-
0.
|
190 |
-
0.
|
191 |
-
0.007838719844818116,
|
192 |
-
0.007895040035247802,
|
193 |
-
0.007795711994171142,
|
194 |
-
0.007676959991455078,
|
195 |
-
0.007639039993286132,
|
196 |
-
0.007779327869415284,
|
197 |
-
0.00764518404006958,
|
198 |
-
0.007531455993652344,
|
199 |
-
0.007616511821746826,
|
200 |
0.0075673599243164065,
|
201 |
-
0.
|
202 |
-
0.
|
203 |
-
0.
|
204 |
-
0.
|
205 |
-
0.
|
206 |
-
0.
|
207 |
-
0.
|
208 |
-
0.
|
209 |
-
0.
|
210 |
-
0.
|
211 |
-
0.
|
212 |
-
0.
|
213 |
-
0.
|
214 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
215 |
0.00707583999633789,
|
216 |
-
0.0070348801612854,
|
217 |
-
0.006974431991577148,
|
218 |
-
0.006982656002044678,
|
219 |
-
0.007017471790313721,
|
220 |
-
0.0069847040176391605,
|
221 |
-
0.00698367977142334,
|
222 |
-
0.007009280204772949,
|
223 |
-
0.00695091199874878,
|
224 |
-
0.006933504104614257,
|
225 |
-
0.006907904148101806,
|
226 |
-
0.0069253120422363285,
|
227 |
-
0.006922304153442383,
|
228 |
-
0.006904895782470703,
|
229 |
-
0.006916096210479736,
|
230 |
-
0.006917119979858399,
|
231 |
-
0.00693452787399292,
|
232 |
-
0.006880256175994873,
|
233 |
-
0.0069253120422363285,
|
234 |
-
0.006924287796020508,
|
235 |
-
0.007024640083312988,
|
236 |
-
0.007002111911773682,
|
237 |
-
0.0070522880554199216,
|
238 |
-
0.0070225920677185055,
|
239 |
-
0.007000063896179199,
|
240 |
-
0.0069847040176391605,
|
241 |
-
0.00701638412475586,
|
242 |
-
0.007005184173583984,
|
243 |
-
0.007016448020935059,
|
244 |
-
0.006951935768127441,
|
245 |
-
0.006957056045532227,
|
246 |
-
0.0069621758460998535,
|
247 |
-
0.006979584217071533,
|
248 |
-
0.0069550080299377445,
|
249 |
-
0.007018496036529541,
|
250 |
-
0.006977536201477051,
|
251 |
-
0.0069918718338012695,
|
252 |
-
0.006967296123504638,
|
253 |
-
0.006957056045532227,
|
254 |
0.007160831928253173,
|
255 |
-
0.
|
|
|
|
|
256 |
]
|
257 |
},
|
258 |
"throughput": {
|
259 |
"unit": "samples/s",
|
260 |
-
"value":
|
261 |
},
|
262 |
"energy": {
|
263 |
"unit": "kWh",
|
264 |
-
"cpu": 8.
|
265 |
-
"ram": 4.
|
266 |
-
"gpu": 1.
|
267 |
-
"total":
|
268 |
},
|
269 |
"efficiency": {
|
270 |
"unit": "samples/kWh",
|
271 |
-
"value":
|
272 |
}
|
273 |
}
|
274 |
}
|
|
|
6 |
"version": "2.3.0+cu121",
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "multiple-choice",
|
|
|
9 |
"library": "transformers",
|
10 |
+
"model": "FacebookAI/roberta-base",
|
11 |
+
"processor": "FacebookAI/roberta-base",
|
12 |
"device": "cuda",
|
13 |
"device_ids": "0",
|
14 |
"seed": 42,
|
15 |
"inter_op_num_threads": null,
|
16 |
"intra_op_num_threads": null,
|
17 |
+
"model_kwargs": {},
|
18 |
+
"processor_kwargs": {},
|
19 |
+
"hub_kwargs": {},
|
|
|
|
|
|
|
20 |
"no_weights": true,
|
21 |
"device_map": null,
|
22 |
"torch_dtype": null,
|
|
|
102 |
"forward": {
|
103 |
"memory": {
|
104 |
"unit": "MB",
|
105 |
+
"max_ram": 909.84448,
|
106 |
"max_global_vram": 1195.900928,
|
107 |
"max_process_vram": 0.0,
|
108 |
"max_reserved": 555.74528,
|
|
|
110 |
},
|
111 |
"latency": {
|
112 |
"unit": "s",
|
113 |
+
"count": 129,
|
114 |
+
"total": 0.9967399663925174,
|
115 |
+
"mean": 0.007726666406143543,
|
116 |
+
"stdev": 0.00044555517191184835,
|
117 |
+
"p50": 0.007923711776733398,
|
118 |
+
"p90": 0.008125644493103027,
|
119 |
+
"p95": 0.00828538875579834,
|
120 |
+
"p99": 0.008613314170837402,
|
121 |
"values": [
|
122 |
+
0.00859340763092041,
|
123 |
+
0.00818892765045166,
|
124 |
+
0.008342528343200683,
|
125 |
+
0.008352736473083495,
|
126 |
+
0.008274944305419921,
|
127 |
+
0.008227840423583984,
|
128 |
+
0.00819916820526123,
|
129 |
+
0.008147968292236327,
|
130 |
+
0.008110079765319824,
|
131 |
+
0.007993311882019042,
|
132 |
+
0.008053759574890136,
|
133 |
+
0.00802406406402588,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
134 |
0.007942143917083741,
|
135 |
+
0.007960576057434082,
|
136 |
+
0.007977983951568603,
|
137 |
+
0.008027135848999023,
|
138 |
+
0.008110079765319824,
|
139 |
+
0.008085503578186035,
|
140 |
+
0.008070143699645996,
|
141 |
+
0.008017919540405273,
|
142 |
+
0.00801375961303711,
|
143 |
+
0.008087552070617676,
|
144 |
+
0.008076288223266602,
|
145 |
+
0.008067071914672852,
|
146 |
+
0.008048640251159669,
|
147 |
+
0.008053759574890136,
|
148 |
+
0.007998464107513427,
|
149 |
+
0.008004544258117675,
|
150 |
+
0.008047616004943848,
|
151 |
+
0.00800153636932373,
|
152 |
+
0.007889920234680176,
|
153 |
+
0.007923711776733398,
|
154 |
+
0.007968768119812012,
|
155 |
+
0.00793497610092163,
|
156 |
+
0.007892960071563721,
|
157 |
+
0.007990272045135497,
|
158 |
+
0.007916543960571289,
|
159 |
+
0.00809062385559082,
|
160 |
+
0.008065024375915527,
|
161 |
+
0.008017919540405273,
|
162 |
+
0.008537088394165039,
|
163 |
+
0.008292351722717285,
|
164 |
+
0.008142848014831543,
|
165 |
+
0.008769536018371582,
|
166 |
+
0.008030207633972167,
|
167 |
+
0.00800972843170166,
|
168 |
+
0.008012800216674805,
|
169 |
0.00795136022567749,
|
170 |
+
0.007938047885894776,
|
171 |
+
0.007922688007354736,
|
172 |
+
0.007935999870300293,
|
173 |
+
0.008059904098510743,
|
174 |
+
0.007944191932678223,
|
175 |
+
0.007882751941680909,
|
176 |
+
0.00791756820678711,
|
177 |
+
0.007979008197784423,
|
178 |
+
0.007970816135406494,
|
179 |
+
0.00793497610092163,
|
180 |
+
0.007947264194488525,
|
181 |
+
0.007994431972503662,
|
182 |
+
0.00808448028564453,
|
183 |
+
0.00808448028564453,
|
184 |
+
0.008010751724243164,
|
185 |
+
0.00809881591796875,
|
186 |
+
0.008038399696350097,
|
187 |
+
0.008002559661865234,
|
188 |
+
0.008621055603027344,
|
189 |
+
0.008121343612670898,
|
190 |
+
0.007927807807922363,
|
191 |
+
0.007854080200195313,
|
192 |
+
0.0077506561279296875,
|
193 |
+
0.007968768119812012,
|
194 |
+
0.007830527782440186,
|
195 |
+
0.007584767818450928,
|
196 |
+
0.007550975799560547,
|
197 |
0.007832575798034667,
|
198 |
+
0.00780185604095459,
|
199 |
+
0.007817215919494629,
|
200 |
+
0.00780291223526001,
|
201 |
+
0.007798783779144287,
|
202 |
+
0.008079360008239746,
|
203 |
+
0.007853055953979492,
|
204 |
+
0.007605247974395752,
|
205 |
+
0.007724031925201416,
|
206 |
+
0.007699456214904785,
|
207 |
+
0.00759500789642334,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
208 |
0.0075673599243164065,
|
209 |
+
0.007521279811859131,
|
210 |
+
0.007562240123748779,
|
211 |
+
0.007494656085968018,
|
212 |
+
0.00739737606048584,
|
213 |
+
0.007401472091674805,
|
214 |
+
0.007328767776489258,
|
215 |
+
0.007106560230255127,
|
216 |
+
0.007146495819091797,
|
217 |
+
0.007126016139984131,
|
218 |
+
0.007114751815795899,
|
219 |
+
0.007124991893768311,
|
220 |
+
0.007079936027526855,
|
221 |
+
0.007107583999633789,
|
222 |
+
0.007081984043121338,
|
223 |
+
0.007159808158874512,
|
224 |
+
0.0070266880989074704,
|
225 |
+
0.00708403205871582,
|
226 |
+
0.00708403205871582,
|
227 |
+
0.007081984043121338,
|
228 |
+
0.00709222412109375,
|
229 |
+
0.007079936027526855,
|
230 |
+
0.0070891518592834475,
|
231 |
+
0.007105535984039306,
|
232 |
+
0.0070830078125,
|
233 |
+
0.007066624164581299,
|
234 |
+
0.007062528133392334,
|
235 |
+
0.0070830078125,
|
236 |
+
0.00709222412109375,
|
237 |
+
0.007076863765716553,
|
238 |
+
0.007069695949554444,
|
239 |
+
0.007106560230255127,
|
240 |
+
0.007131135940551757,
|
241 |
+
0.007095295906066895,
|
242 |
+
0.007061503887176514,
|
243 |
+
0.007078911781311035,
|
244 |
+
0.007087103843688965,
|
245 |
+
0.00709222412109375,
|
246 |
0.00707583999633789,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
247 |
0.007160831928253173,
|
248 |
+
0.007081984043121338,
|
249 |
+
0.007110655784606934,
|
250 |
+
0.0070563840866088865
|
251 |
]
|
252 |
},
|
253 |
"throughput": {
|
254 |
"unit": "samples/s",
|
255 |
+
"value": 129.42191980811944
|
256 |
},
|
257 |
"energy": {
|
258 |
"unit": "kWh",
|
259 |
+
"cpu": 8.951583153621893e-08,
|
260 |
+
"ram": 4.881614143489702e-08,
|
261 |
+
"gpu": 1.6610915544361268e-07,
|
262 |
+
"total": 3.044411284147286e-07
|
263 |
},
|
264 |
"efficiency": {
|
265 |
"unit": "samples/kWh",
|
266 |
+
"value": 3284707.3101034425
|
267 |
}
|
268 |
}
|
269 |
}
|