IlyasMoutawwakil HF staff commited on
Commit
939454f
·
verified ·
1 Parent(s): e17172c

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -6,19 +6,17 @@
6
  "version": "2.3.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
- "model": "FacebookAI/roberta-base",
10
  "library": "transformers",
 
 
11
  "device": "cuda",
12
  "device_ids": "0",
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
- "hub_kwargs": {
17
- "revision": "main",
18
- "force_download": false,
19
- "local_files_only": false,
20
- "trust_remote_code": false
21
- },
22
  "no_weights": true,
23
  "device_map": null,
24
  "torch_dtype": null,
@@ -104,7 +102,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 909.647872,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -112,163 +110,160 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 132,
116
- "total": 1.0009157433509823,
117
- "mean": 0.0075826950253862335,
118
- "stdev": 0.0004717101041693446,
119
- "p50": 0.0077972478866577145,
120
- "p90": 0.008067788982391359,
121
- "p95": 0.008177561378479003,
122
- "p99": 0.008570173263549804,
123
  "values": [
124
- 0.008353792190551757,
125
- 0.008102911949157715,
126
- 0.008176639556884765,
127
- 0.00820633602142334,
128
- 0.00811520004272461,
129
- 0.008766464233398438,
130
- 0.008263680458068847,
131
- 0.008094719886779785,
132
- 0.008137727737426758,
133
- 0.008072192192077637,
134
- 0.008028160095214844,
135
- 0.007994368076324462,
136
- 0.0078919677734375,
137
- 0.007888895988464355,
138
- 0.008018943786621094,
139
- 0.007961599826812745,
140
- 0.008143872261047362,
141
- 0.007972864151000977,
142
- 0.007878623962402343,
143
- 0.00786636781692505,
144
- 0.007832575798034667,
145
- 0.007867392063140868,
146
- 0.007902207851409913,
147
- 0.007871488094329833,
148
- 0.007935999870300293,
149
- 0.007916543960571289,
150
- 0.007854080200195313,
151
- 0.007862271785736084,
152
- 0.007865344047546387,
153
- 0.00787660789489746,
154
- 0.007796703815460205,
155
- 0.00773529577255249,
156
- 0.0077761921882629394,
157
- 0.007821311950683594,
158
- 0.007766016006469726,
159
- 0.007824384212493896,
160
- 0.007837696075439453,
161
- 0.007839744091033935,
162
  0.007942143917083741,
163
- 0.007881728172302246,
164
- 0.00790835189819336,
165
- 0.007870463848114014,
166
- 0.007803904056549072,
167
- 0.007863296031951903,
168
- 0.00788479995727539,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
169
  0.00795136022567749,
170
- 0.007874559879302979,
171
- 0.00786636781692505,
172
- 0.00860159969329834,
173
- 0.007869440078735352,
174
- 0.007797760009765625,
175
- 0.007833600044250488,
176
- 0.007899136066436767,
177
- 0.007844863891601562,
178
- 0.007774208068847656,
179
- 0.007836607933044434,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
180
  0.007832575798034667,
181
- 0.007808000087738037,
182
- 0.00783564805984497,
183
- 0.007796735763549805,
184
- 0.008178688049316407,
185
- 0.007988224029541016,
186
- 0.007955455780029297,
187
- 0.007918591976165772,
188
- 0.007870463848114014,
189
- 0.007869440078735352,
190
- 0.008500224113464355,
191
- 0.007838719844818116,
192
- 0.007895040035247802,
193
- 0.007795711994171142,
194
- 0.007676959991455078,
195
- 0.007639039993286132,
196
- 0.007779327869415284,
197
- 0.00764518404006958,
198
- 0.007531455993652344,
199
- 0.007616511821746826,
200
  0.0075673599243164065,
201
- 0.007725056171417236,
202
- 0.007749631881713868,
203
- 0.007756800174713135,
204
- 0.007753727912902832,
205
- 0.008012800216674805,
206
- 0.007870463848114014,
207
- 0.007673855781555176,
208
- 0.007953407764434815,
209
- 0.007183328151702881,
210
- 0.007005184173583984,
211
- 0.007013376235961914,
212
- 0.006957056045532227,
213
- 0.006985727787017822,
214
- 0.007013376235961914,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
215
  0.00707583999633789,
216
- 0.0070348801612854,
217
- 0.006974431991577148,
218
- 0.006982656002044678,
219
- 0.007017471790313721,
220
- 0.0069847040176391605,
221
- 0.00698367977142334,
222
- 0.007009280204772949,
223
- 0.00695091199874878,
224
- 0.006933504104614257,
225
- 0.006907904148101806,
226
- 0.0069253120422363285,
227
- 0.006922304153442383,
228
- 0.006904895782470703,
229
- 0.006916096210479736,
230
- 0.006917119979858399,
231
- 0.00693452787399292,
232
- 0.006880256175994873,
233
- 0.0069253120422363285,
234
- 0.006924287796020508,
235
- 0.007024640083312988,
236
- 0.007002111911773682,
237
- 0.0070522880554199216,
238
- 0.0070225920677185055,
239
- 0.007000063896179199,
240
- 0.0069847040176391605,
241
- 0.00701638412475586,
242
- 0.007005184173583984,
243
- 0.007016448020935059,
244
- 0.006951935768127441,
245
- 0.006957056045532227,
246
- 0.0069621758460998535,
247
- 0.006979584217071533,
248
- 0.0069550080299377445,
249
- 0.007018496036529541,
250
- 0.006977536201477051,
251
- 0.0069918718338012695,
252
- 0.006967296123504638,
253
- 0.006957056045532227,
254
  0.007160831928253173,
255
- 0.007192575931549072
 
 
256
  ]
257
  },
258
  "throughput": {
259
  "unit": "samples/s",
260
- "value": 131.87923246973313
261
  },
262
  "energy": {
263
  "unit": "kWh",
264
- "cpu": 8.439567729748545e-08,
265
- "ram": 4.6138346856713856e-08,
266
- "gpu": 1.5222628411346766e-07,
267
- "total": 2.82760308267667e-07
268
  },
269
  "efficiency": {
270
  "unit": "samples/kWh",
271
- "value": 3536564.2587055694
272
  }
273
  }
274
  }
 
6
  "version": "2.3.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
 
9
  "library": "transformers",
10
+ "model": "FacebookAI/roberta-base",
11
+ "processor": "FacebookAI/roberta-base",
12
  "device": "cuda",
13
  "device_ids": "0",
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
 
 
 
20
  "no_weights": true,
21
  "device_map": null,
22
  "torch_dtype": null,
 
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 909.84448,
106
  "max_global_vram": 1195.900928,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 555.74528,
 
110
  },
111
  "latency": {
112
  "unit": "s",
113
+ "count": 129,
114
+ "total": 0.9967399663925174,
115
+ "mean": 0.007726666406143543,
116
+ "stdev": 0.00044555517191184835,
117
+ "p50": 0.007923711776733398,
118
+ "p90": 0.008125644493103027,
119
+ "p95": 0.00828538875579834,
120
+ "p99": 0.008613314170837402,
121
  "values": [
122
+ 0.00859340763092041,
123
+ 0.00818892765045166,
124
+ 0.008342528343200683,
125
+ 0.008352736473083495,
126
+ 0.008274944305419921,
127
+ 0.008227840423583984,
128
+ 0.00819916820526123,
129
+ 0.008147968292236327,
130
+ 0.008110079765319824,
131
+ 0.007993311882019042,
132
+ 0.008053759574890136,
133
+ 0.00802406406402588,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
134
  0.007942143917083741,
135
+ 0.007960576057434082,
136
+ 0.007977983951568603,
137
+ 0.008027135848999023,
138
+ 0.008110079765319824,
139
+ 0.008085503578186035,
140
+ 0.008070143699645996,
141
+ 0.008017919540405273,
142
+ 0.00801375961303711,
143
+ 0.008087552070617676,
144
+ 0.008076288223266602,
145
+ 0.008067071914672852,
146
+ 0.008048640251159669,
147
+ 0.008053759574890136,
148
+ 0.007998464107513427,
149
+ 0.008004544258117675,
150
+ 0.008047616004943848,
151
+ 0.00800153636932373,
152
+ 0.007889920234680176,
153
+ 0.007923711776733398,
154
+ 0.007968768119812012,
155
+ 0.00793497610092163,
156
+ 0.007892960071563721,
157
+ 0.007990272045135497,
158
+ 0.007916543960571289,
159
+ 0.00809062385559082,
160
+ 0.008065024375915527,
161
+ 0.008017919540405273,
162
+ 0.008537088394165039,
163
+ 0.008292351722717285,
164
+ 0.008142848014831543,
165
+ 0.008769536018371582,
166
+ 0.008030207633972167,
167
+ 0.00800972843170166,
168
+ 0.008012800216674805,
169
  0.00795136022567749,
170
+ 0.007938047885894776,
171
+ 0.007922688007354736,
172
+ 0.007935999870300293,
173
+ 0.008059904098510743,
174
+ 0.007944191932678223,
175
+ 0.007882751941680909,
176
+ 0.00791756820678711,
177
+ 0.007979008197784423,
178
+ 0.007970816135406494,
179
+ 0.00793497610092163,
180
+ 0.007947264194488525,
181
+ 0.007994431972503662,
182
+ 0.00808448028564453,
183
+ 0.00808448028564453,
184
+ 0.008010751724243164,
185
+ 0.00809881591796875,
186
+ 0.008038399696350097,
187
+ 0.008002559661865234,
188
+ 0.008621055603027344,
189
+ 0.008121343612670898,
190
+ 0.007927807807922363,
191
+ 0.007854080200195313,
192
+ 0.0077506561279296875,
193
+ 0.007968768119812012,
194
+ 0.007830527782440186,
195
+ 0.007584767818450928,
196
+ 0.007550975799560547,
197
  0.007832575798034667,
198
+ 0.00780185604095459,
199
+ 0.007817215919494629,
200
+ 0.00780291223526001,
201
+ 0.007798783779144287,
202
+ 0.008079360008239746,
203
+ 0.007853055953979492,
204
+ 0.007605247974395752,
205
+ 0.007724031925201416,
206
+ 0.007699456214904785,
207
+ 0.00759500789642334,
 
 
 
 
 
 
 
 
 
208
  0.0075673599243164065,
209
+ 0.007521279811859131,
210
+ 0.007562240123748779,
211
+ 0.007494656085968018,
212
+ 0.00739737606048584,
213
+ 0.007401472091674805,
214
+ 0.007328767776489258,
215
+ 0.007106560230255127,
216
+ 0.007146495819091797,
217
+ 0.007126016139984131,
218
+ 0.007114751815795899,
219
+ 0.007124991893768311,
220
+ 0.007079936027526855,
221
+ 0.007107583999633789,
222
+ 0.007081984043121338,
223
+ 0.007159808158874512,
224
+ 0.0070266880989074704,
225
+ 0.00708403205871582,
226
+ 0.00708403205871582,
227
+ 0.007081984043121338,
228
+ 0.00709222412109375,
229
+ 0.007079936027526855,
230
+ 0.0070891518592834475,
231
+ 0.007105535984039306,
232
+ 0.0070830078125,
233
+ 0.007066624164581299,
234
+ 0.007062528133392334,
235
+ 0.0070830078125,
236
+ 0.00709222412109375,
237
+ 0.007076863765716553,
238
+ 0.007069695949554444,
239
+ 0.007106560230255127,
240
+ 0.007131135940551757,
241
+ 0.007095295906066895,
242
+ 0.007061503887176514,
243
+ 0.007078911781311035,
244
+ 0.007087103843688965,
245
+ 0.00709222412109375,
246
  0.00707583999633789,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
247
  0.007160831928253173,
248
+ 0.007081984043121338,
249
+ 0.007110655784606934,
250
+ 0.0070563840866088865
251
  ]
252
  },
253
  "throughput": {
254
  "unit": "samples/s",
255
+ "value": 129.42191980811944
256
  },
257
  "energy": {
258
  "unit": "kWh",
259
+ "cpu": 8.951583153621893e-08,
260
+ "ram": 4.881614143489702e-08,
261
+ "gpu": 1.6610915544361268e-07,
262
+ "total": 3.044411284147286e-07
263
  },
264
  "efficiency": {
265
  "unit": "samples/kWh",
266
+ "value": 3284707.3101034425
267
  }
268
  }
269
  }