IlyasMoutawwakil HF staff commited on
Commit
5d96e2d
·
verified ·
1 Parent(s): 503149e

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.1+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -76,7 +76,7 @@
76
  "cpu_ram_mb": 66697.261056,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
@@ -88,9 +88,9 @@
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.44.2",
90
  "transformers_commit": null,
91
- "accelerate_version": "0.34.0",
92
  "accelerate_commit": null,
93
- "diffusers_version": "0.30.2",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
@@ -104,8 +104,8 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 799.834112,
108
- "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.37152
@@ -113,195 +113,193 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.8667509765625,
117
- "mean": 7.8667509765625,
118
  "stdev": 0.0,
119
- "p50": 7.8667509765625,
120
- "p90": 7.8667509765625,
121
- "p95": 7.8667509765625,
122
- "p99": 7.8667509765625,
123
  "values": [
124
- 7.8667509765625
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 9.238619541668985e-07,
131
- "ram": 4.6324608645141225e-07,
132
- "gpu": 1.548056794000146e-06,
133
- "total": 2.935164834618457e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 1096.564736,
141
- "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
144
  "max_allocated": 509.038592
145
  },
146
  "latency": {
147
  "unit": "s",
148
- "count": 132,
149
- "total": 1.002677695274353,
150
- "mean": 0.007596043146017826,
151
- "stdev": 0.0004537093049689842,
152
- "p50": 0.007549952030181885,
153
- "p90": 0.00763105297088623,
154
- "p95": 0.007710315227508545,
155
- "p99": 0.010351206302642822,
156
  "values": [
157
- 0.010590208053588868,
158
- 0.010363903999328614,
159
- 0.010322943687438964,
160
- 0.008053759574890136,
161
- 0.00770249605178833,
162
- 0.007675903797149658,
163
- 0.007605247974395752,
164
- 0.008141823768615723,
165
- 0.007363552093505859,
166
- 0.007332863807678222,
167
- 0.007366591930389404,
168
- 0.007448575973510742,
169
- 0.007341055870056152,
170
- 0.007160799980163574,
171
- 0.007520256042480469,
172
- 0.007193600177764893,
173
- 0.007187456130981445,
174
- 0.0071905279159545895,
175
- 0.007120895862579346,
176
- 0.007132160186767578,
177
- 0.007173120021820068,
178
- 0.007227392196655274,
179
- 0.0071485438346862796,
180
- 0.00714137601852417,
181
- 0.007576608180999756,
182
- 0.007681951999664306,
183
- 0.007584767818450928,
184
  0.007606272220611572,
185
- 0.0075980801582336424,
186
- 0.007600128173828125,
187
- 0.007589888095855713,
188
- 0.00759603214263916,
189
- 0.007520256042480469,
190
- 0.007556096076965332,
191
- 0.007549952030181885,
192
- 0.007599103927612305,
193
- 0.007550975799560547,
194
- 0.007575551986694336,
195
- 0.007549952030181885,
196
- 0.007549952030181885,
197
- 0.007558144092559814,
198
- 0.007638016223907471,
199
- 0.007510975837707519,
200
- 0.007574528217315674,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
  0.007565311908721924,
202
- 0.007550975799560547,
203
- 0.007559167861938477,
204
- 0.007553023815155029,
205
- 0.007541759967803955,
206
- 0.007555071830749512,
 
 
207
  0.007539711952209473,
208
- 0.007517183780670166,
209
- 0.007542784214019775,
210
- 0.007552000045776367,
211
  0.007529471874237061,
212
- 0.007574495792388916,
213
- 0.007550975799560547,
214
- 0.007566336154937744,
215
- 0.0075673599243164065,
216
- 0.0075673599243164065,
217
- 0.007579648017883301,
218
- 0.007631872177124023,
219
- 0.007623680114746094,
 
 
220
  0.007585792064666748,
221
- 0.007562240123748779,
222
- 0.00753766393661499,
223
- 0.007593023777008057,
224
- 0.007639039993286132,
225
- 0.007577600002288819,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
226
  0.0075673599243164065,
227
- 0.007525375843048096,
228
- 0.007621632099151611,
229
- 0.0075632638931274416,
230
  0.007544832229614258,
231
- 0.007572480201721191,
232
- 0.007599103927612305,
233
- 0.007572415828704834,
234
- 0.007574528217315674,
235
- 0.00753868818283081,
236
- 0.007573503971099854,
237
- 0.0075560641288757325,
238
- 0.00753868818283081,
239
- 0.007582719802856445,
240
- 0.007565311908721924,
241
- 0.007527423858642578,
242
- 0.007535615921020508,
243
- 0.007528384208679199,
244
- 0.007535615921020508,
245
- 0.007519167900085449,
246
- 0.007561215877532959,
247
- 0.007506944179534912,
248
- 0.007719871997833252,
249
- 0.007556096076965332,
250
- 0.007559167861938477,
251
- 0.007511040210723877,
252
- 0.007656447887420655,
253
- 0.007552000045776367,
254
- 0.007619584083557129,
255
- 0.007526400089263916,
256
- 0.007524352073669433,
257
- 0.007493631839752197,
258
- 0.007531519889831543,
259
- 0.007502848148345947,
260
- 0.007560192108154297,
261
- 0.007525375843048096,
262
- 0.007532544136047363,
263
- 0.007535615921020508,
264
- 0.00753868818283081,
265
- 0.007555071830749512,
266
- 0.007520256042480469,
267
  0.00749567985534668,
268
- 0.007507904052734375,
269
- 0.0074997758865356446,
270
- 0.0075008001327514645,
271
- 0.007502848148345947,
272
- 0.007555071830749512,
273
- 0.0075049281120300295,
274
- 0.007526400089263916,
275
- 0.007478271961212158,
276
- 0.007517183780670166,
277
- 0.007522304058074952,
278
- 0.007518208026885987,
279
- 0.007507967948913574,
280
- 0.007531519889831543,
281
- 0.007520256042480469,
282
- 0.007546879768371582,
283
- 0.007518208026885987,
284
- 0.0075008001327514645,
285
- 0.007526400089263916,
286
- 0.007521279811859131,
287
- 0.0074926080703735356,
288
- 0.007915520191192627
289
  ]
290
  },
291
  "throughput": {
292
  "unit": "samples/s",
293
- "value": 131.64748814311872
294
  },
295
  "energy": {
296
  "unit": "kWh",
297
- "cpu": 8.877624120542962e-08,
298
- "ram": 4.854012163597355e-08,
299
- "gpu": 1.5776339320895888e-07,
300
- "total": 2.9507975605036207e-07
301
  },
302
  "efficiency": {
303
  "unit": "samples/kWh",
304
- "value": 3388914.283327953
305
  }
306
  }
307
  }
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.6.0.dev20240917+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
76
  "cpu_ram_mb": 66697.261056,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
 
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.44.2",
90
  "transformers_commit": null,
91
+ "accelerate_version": "0.34.2",
92
  "accelerate_commit": null,
93
+ "diffusers_version": "0.30.3",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 810.094592,
108
+ "max_global_vram": 1192.7552,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.37152
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 8.5225263671875,
117
+ "mean": 8.5225263671875,
118
  "stdev": 0.0,
119
+ "p50": 8.5225263671875,
120
+ "p90": 8.5225263671875,
121
+ "p95": 8.5225263671875,
122
+ "p99": 8.5225263671875,
123
  "values": [
124
+ 8.5225263671875
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 9.056904298379071e-07,
131
+ "ram": 4.69438917347047e-07,
132
+ "gpu": 0.0,
133
+ "total": 1.3751293471849541e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 1103.31904,
141
+ "max_global_vram": 1203.24096,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
144
  "max_allocated": 509.038592
145
  },
146
  "latency": {
147
  "unit": "s",
148
+ "count": 130,
149
+ "total": 0.9992434840202338,
150
+ "mean": 0.007686488338617178,
151
+ "stdev": 0.00032106215549831127,
152
+ "p50": 0.007704576015472412,
153
+ "p90": 0.007926271820068359,
154
+ "p95": 0.008019609260559082,
155
+ "p99": 0.00890674160003662,
156
  "values": [
157
+ 0.007815167903900147,
158
+ 0.007880703926086426,
159
+ 0.007830527782440186,
160
+ 0.007838719844818116,
161
+ 0.007809023857116699,
162
+ 0.007702527999877929,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
163
  0.007606272220611572,
164
+ 0.007706624031066894,
165
+ 0.007636991977691651,
166
+ 0.007614463806152344,
167
+ 0.007639039993286132,
168
+ 0.007628799915313721,
169
+ 0.007648255825042725,
170
+ 0.007611328125,
171
+ 0.007732223987579345,
172
+ 0.007844863891601562,
173
+ 0.007861248016357422,
174
+ 0.007894015789031983,
175
+ 0.007941120147705078,
176
+ 0.00791756820678711,
177
+ 0.007921664237976075,
178
+ 0.007913472175598145,
179
+ 0.00785100793838501,
180
+ 0.007865344047546387,
181
+ 0.007903232097625732,
182
+ 0.007930880069732665,
183
+ 0.007887872219085693,
184
+ 0.007910399913787843,
185
+ 0.007914495944976807,
186
+ 0.007910399913787843,
187
+ 0.007979008197784423,
188
+ 0.00790835189819336,
189
+ 0.00790118408203125,
190
+ 0.007910399913787843,
191
+ 0.007907328128814697,
192
+ 0.00785097599029541,
193
+ 0.00783564805984497,
194
+ 0.007825407981872558,
195
+ 0.008006655693054199,
196
+ 0.007886847972869874,
197
+ 0.007844863891601562,
198
+ 0.00784281587600708,
199
+ 0.007839744091033935,
200
+ 0.007833600044250488,
201
+ 0.007882751941680909,
202
+ 0.008030207633972167,
203
+ 0.008048640251159669,
204
+ 0.007914495944976807,
205
+ 0.007879680156707763,
206
+ 0.007867392063140868,
207
+ 0.007875584125518798,
208
+ 0.007830527782440186,
209
+ 0.007902207851409913,
210
+ 0.00790937614440918,
211
+ 0.007840767860412597,
212
+ 0.007860223770141601,
213
+ 0.007925759792327881,
214
+ 0.007777279853820801,
215
+ 0.007763967990875244,
216
+ 0.007840767860412597,
217
+ 0.007514111995697022,
218
  0.007565311908721924,
219
+ 0.008408063888549805,
220
+ 0.007550943851470948,
221
+ 0.007501823902130127,
222
+ 0.007584767818450928,
223
+ 0.007523327827453613,
224
+ 0.007540736198425293,
225
+ 0.007548927783966064,
226
  0.007539711952209473,
 
 
 
227
  0.007529471874237061,
228
+ 0.007797760009765625,
229
+ 0.007933951854705811,
230
+ 0.007975935935974121,
231
+ 0.0076267518997192385,
232
+ 0.007618559837341309,
233
+ 0.00764518404006958,
234
+ 0.007601151943206787,
235
+ 0.007624703884124756,
236
+ 0.007607295989990235,
237
+ 0.007527391910552978,
238
  0.007585792064666748,
239
+ 0.007606272220611572,
240
+ 0.007504896163940429,
241
+ 0.007490560054779053,
242
+ 0.0073686399459838865,
243
+ 0.0074065918922424315,
244
+ 0.007481344223022461,
245
+ 0.007325695991516113,
246
+ 0.007203839778900147,
247
+ 0.007305215835571289,
248
+ 0.007715839862823487,
249
+ 0.007786496162414551,
250
+ 0.00784281587600708,
251
+ 0.007873536109924317,
252
+ 0.008051712036132813,
253
+ 0.007911424160003662,
254
+ 0.008927231788635253,
255
+ 0.008856575965881347,
256
+ 0.008948736190795899,
257
+ 0.0074414081573486324,
258
+ 0.007374847888946533,
259
+ 0.0073471999168395995,
260
+ 0.007342080116271973,
261
+ 0.007305215835571289,
262
+ 0.007265279769897461,
263
+ 0.007205887794494629,
264
+ 0.007296000003814697,
265
+ 0.00724070405960083,
266
+ 0.0071905279159545895,
267
+ 0.007176191806793213,
268
+ 0.007172095775604248,
269
+ 0.007180287837982178,
270
+ 0.007186431884765625,
271
+ 0.007169023990631103,
272
+ 0.007177216053009033,
273
+ 0.0071782078742980955,
274
+ 0.007178239822387696,
275
+ 0.007189504146575928,
276
+ 0.007207935810089112,
277
+ 0.007172095775604248,
278
+ 0.007194623947143554,
279
+ 0.0075642881393432615,
280
  0.0075673599243164065,
281
+ 0.007569407939910889,
282
+ 0.007583712100982666,
 
283
  0.007544832229614258,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
284
  0.00749567985534668,
285
+ 0.0075345921516418455,
286
+ 0.007542784214019775
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
287
  ]
288
  },
289
  "throughput": {
290
  "unit": "samples/s",
291
+ "value": 130.09842153483356
292
  },
293
  "energy": {
294
  "unit": "kWh",
295
+ "cpu": 8.885693587796052e-08,
296
+ "ram": 4.8445208733740644e-08,
297
+ "gpu": 1.6774889041793622e-07,
298
+ "total": 3.0505103502963736e-07
299
  },
300
  "efficiency": {
301
  "unit": "samples/kWh",
302
+ "value": 3278140.0000916063
303
  }
304
  }
305
  }