IlyasMoutawwakil HF staff commited on
Commit
e5a36aa
·
verified ·
1 Parent(s): 39cb52a

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -73,7 +73,7 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7R32",
75
  "cpu_count": 16,
76
- "cpu_ram_mb": 66697.29792,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
  "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
@@ -104,7 +104,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 909.94688,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -112,167 +112,160 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 136,
116
- "total": 0.9998827495574946,
117
- "mean": 0.007352079040863935,
118
- "stdev": 0.00022782686304659592,
119
- "p50": 0.007419903993606567,
120
- "p90": 0.00756224012374878,
121
- "p95": 0.007704320192337036,
122
- "p99": 0.00813306851387024,
123
  "values": [
124
- 0.00819206428527832,
125
- 0.007654399871826172,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
126
  0.007823359966278077,
127
- 0.007812096118927002,
128
- 0.007699456214904785,
129
- 0.007727104187011719,
130
- 0.007718912124633789,
131
- 0.008173567771911621,
132
- 0.007375807762145996,
133
- 0.007185408115386963,
134
- 0.007161856174468994,
135
- 0.00714035177230835,
136
- 0.007145472049713135,
137
- 0.007138271808624267,
138
- 0.007143424034118652,
139
- 0.007154687881469727,
140
- 0.007142399787902832,
141
- 0.007111680030822754,
142
- 0.007111743927001953,
143
- 0.007129087924957276,
144
- 0.00734822416305542,
145
- 0.007161856174468994,
146
- 0.007121920108795166,
147
- 0.007105472087860107,
148
- 0.007142399787902832,
149
- 0.007129087924957276,
150
- 0.007098368167877197,
151
- 0.007085055828094483,
152
- 0.007114751815795899,
153
- 0.007113728046417236,
154
- 0.007106560230255127,
155
- 0.007131135940551757,
156
- 0.007096320152282715,
157
- 0.00714137601852417,
158
- 0.007126016139984131,
159
- 0.0071188478469848635,
160
- 0.007112736225128174,
161
- 0.00709939193725586,
162
- 0.007113728046417236,
163
- 0.00709222412109375,
164
- 0.007178239822387696,
165
- 0.007219264030456543,
166
- 0.007124991893768311,
167
- 0.007120895862579346,
168
- 0.007120895862579346,
169
- 0.007107583999633789,
170
- 0.00707583999633789,
171
- 0.007108607769012451,
172
- 0.007097343921661377,
173
- 0.007114751815795899,
174
- 0.007105535984039306,
175
- 0.007078879833221435,
176
- 0.007098368167877197,
177
- 0.0071064958572387696,
178
- 0.007119840145111084,
179
- 0.007066624164581299,
180
- 0.007096320152282715,
181
- 0.007101439952850342,
182
- 0.007101439952850342,
183
- 0.007102464199066162,
184
- 0.007071743965148926,
185
- 0.007109632015228271,
186
- 0.007096320152282715,
187
- 0.007104479789733887,
188
- 0.0070522880554199216,
189
  0.007526400089263916,
190
- 0.008057855606079101,
191
- 0.007436287879943848,
192
- 0.00743017578125,
193
- 0.0074414401054382325,
 
 
 
 
 
 
194
  0.007434239864349365,
195
- 0.007445536136627197,
196
- 0.007447487831115722,
197
- 0.007403456211090088,
198
- 0.007459807872772217,
199
- 0.00743833589553833,
200
- 0.0074629120826721195,
201
- 0.007445504188537597,
202
- 0.007426047801971435,
203
- 0.007455743789672851,
204
- 0.007641088008880615,
205
- 0.007463935852050781,
206
- 0.00745472002029419,
207
- 0.007465983867645264,
208
- 0.007445504188537597,
209
- 0.007466047763824463,
210
- 0.007443456172943115,
211
- 0.007445504188537597,
212
- 0.00742195177078247,
213
  0.007409664154052734,
214
- 0.007456768035888672,
 
 
 
 
 
 
 
 
 
 
 
215
  0.007428095817565918,
216
- 0.007615488052368164,
217
- 0.0074301438331604,
 
 
 
218
  0.007442431926727295,
219
- 0.007453695774078369,
220
- 0.007398399829864502,
221
- 0.007444479942321777,
222
- 0.007479296207427978,
223
- 0.0074711041450500485,
224
- 0.007526400089263916,
225
- 0.007524384021759033,
226
- 0.007508992195129394,
227
- 0.007491583824157715,
228
- 0.0074997758865356446,
229
- 0.007515135765075683,
230
- 0.00760319995880127,
231
- 0.007518208026885987,
232
- 0.007579648017883301,
233
- 0.007493631839752197,
234
- 0.007494656085968018,
235
- 0.007523392200469971,
236
- 0.007515135765075683,
237
- 0.007480319976806641,
238
- 0.007473152160644531,
239
- 0.007501823902130127,
240
- 0.007497727870941162,
241
- 0.007574528217315674,
242
- 0.0075335679054260255,
243
- 0.007549952030181885,
244
- 0.007526400089263916,
245
- 0.00749567985534668,
246
- 0.007483391761779785,
247
- 0.007417856216430664,
248
- 0.007434239864349365,
249
- 0.007422976016998291,
250
- 0.007408639907836914,
251
- 0.00738099193572998,
252
  0.007401472091674805,
253
- 0.007372799873352051,
254
- 0.007361536026000977,
255
- 0.007404543876647949,
256
- 0.007368703842163086,
257
- 0.007461887836456299,
258
- 0.007478271961212158,
259
- 0.007488512039184571
260
  ]
261
  },
262
  "throughput": {
263
  "unit": "samples/s",
264
- "value": 136.01594793008252
265
  },
266
  "energy": {
267
  "unit": "kWh",
268
- "cpu": 8.692752149950076e-08,
269
- "ram": 4.751775817381368e-08,
270
- "gpu": 1.5732169925547403e-07,
271
- "total": 2.917669789287885e-07
272
  },
273
  "efficiency": {
274
  "unit": "samples/kWh",
275
- "value": 3427392.6531078415
276
  }
277
  }
278
  }
 
73
  "environment": {
74
  "cpu": " AMD EPYC 7R32",
75
  "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.289728,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
  "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 910.188544,
108
  "max_global_vram": 1195.900928,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 129,
116
+ "total": 0.9989077768325808,
117
+ "mean": 0.00774347113823706,
118
+ "stdev": 0.0002607245609967047,
119
+ "p50": 0.00783564805984497,
120
+ "p90": 0.008009728240966798,
121
+ "p95": 0.008088575744628906,
122
+ "p99": 0.008236892280578613,
123
  "values": [
124
+ 0.007982079982757568,
125
+ 0.00809062385559082,
126
+ 0.008220671653747558,
127
+ 0.008196096420288086,
128
+ 0.008102911949157715,
129
+ 0.008243200302124023,
130
+ 0.008034303665161132,
131
+ 0.007986176013946533,
132
+ 0.007928832054138184,
133
+ 0.00790937614440918,
134
+ 0.007933951854705811,
135
+ 0.007903232097625732,
136
+ 0.007856095790863037,
137
+ 0.007954432010650634,
138
+ 0.007896063804626464,
139
+ 0.007915520191192627,
140
+ 0.007939072132110595,
141
+ 0.008023039817810058,
142
+ 0.00800153636932373,
143
+ 0.007932928085327149,
144
+ 0.007906303882598878,
145
+ 0.007937024116516114,
146
+ 0.008002559661865234,
147
+ 0.00800870418548584,
148
+ 0.007970816135406494,
149
+ 0.007996416091918946,
150
+ 0.007996416091918946,
151
+ 0.007903232097625732,
152
+ 0.007923711776733398,
153
+ 0.007921664237976075,
154
+ 0.00789299201965332,
155
+ 0.00783564805984497,
156
+ 0.00783564805984497,
157
+ 0.007847936153411865,
158
+ 0.007841824054718017,
159
+ 0.007820288181304931,
160
+ 0.007848959922790527,
161
+ 0.007918591976165772,
162
+ 0.008013824462890624,
163
+ 0.008041472434997558,
164
+ 0.00799129581451416,
165
+ 0.007912447929382324,
166
+ 0.007875584125518798,
167
+ 0.00786636781692505,
168
+ 0.007935999870300293,
169
+ 0.007907328128814697,
170
+ 0.007925759792327881,
171
+ 0.007929855823516846,
172
+ 0.007797760009765625,
173
+ 0.007878655910491944,
174
+ 0.007822368144989014,
175
+ 0.007976960182189942,
176
+ 0.007949312210083008,
177
+ 0.007794688224792481,
178
  0.007823359966278077,
179
+ 0.007870368003845215,
180
+ 0.007855103969573975,
181
+ 0.007827455997467042,
182
+ 0.007837696075439453,
183
+ 0.007921664237976075,
184
+ 0.00799232006072998,
185
+ 0.00800051212310791,
186
+ 0.007960576057434082,
187
+ 0.007940095901489258,
188
+ 0.007932928085327149,
189
+ 0.008085503578186035,
190
+ 0.008479743957519532,
191
+ 0.007902207851409913,
192
+ 0.007872511863708496,
193
+ 0.007745535850524903,
194
+ 0.007635968208312988,
195
+ 0.0076574721336364745,
196
+ 0.007724031925201416,
197
+ 0.007580671787261963,
198
+ 0.007483391761779785,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
199
  0.007526400089263916,
200
+ 0.007767039775848389,
201
+ 0.007679999828338623,
202
+ 0.007742464065551758,
203
+ 0.007789567947387695,
204
+ 0.007823359966278077,
205
+ 0.008058879852294922,
206
+ 0.007738368034362793,
207
+ 0.008208383560180664,
208
+ 0.007683072090148926,
209
+ 0.007469056129455566,
210
  0.007434239864349365,
211
+ 0.00738099193572998,
212
+ 0.007418879985809326,
213
+ 0.007425024032592774,
214
+ 0.007411712169647216,
215
+ 0.007444479942321777,
216
+ 0.007458816051483155,
217
+ 0.007417856216430664,
218
+ 0.00743936014175415,
 
 
 
 
 
 
 
 
 
 
219
  0.007409664154052734,
220
+ 0.00738918399810791,
221
+ 0.007419904232025146,
222
+ 0.00742195177078247,
223
+ 0.0074414081573486324,
224
+ 0.00740556812286377,
225
+ 0.0074065918922424315,
226
+ 0.007402495861053467,
227
+ 0.007420928001403809,
228
+ 0.007426047801971435,
229
+ 0.007444479942321777,
230
+ 0.00738918399810791,
231
+ 0.0074291200637817386,
232
  0.007428095817565918,
233
+ 0.007388160228729248,
234
+ 0.007412735939025879,
235
+ 0.00743833589553833,
236
+ 0.007435264110565186,
237
+ 0.007412735939025879,
238
  0.007442431926727295,
239
+ 0.007435264110565186,
240
+ 0.007390207767486572,
241
+ 0.007418879985809326,
242
+ 0.007401472091674805,
243
+ 0.0074035201072692874,
244
+ 0.007426047801971435,
245
+ 0.007401472091674805,
246
+ 0.007411712169647216,
247
+ 0.007395328044891358,
248
+ 0.007436287879943848,
249
+ 0.007623680114746094,
250
+ 0.007485439777374267,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
251
  0.007401472091674805,
252
+ 0.0073870720863342286
 
 
 
 
 
 
253
  ]
254
  },
255
  "throughput": {
256
  "unit": "samples/s",
257
+ "value": 129.1410508476006
258
  },
259
  "energy": {
260
  "unit": "kWh",
261
+ "cpu": 8.705616411235599e-08,
262
+ "ram": 4.758963447407894e-08,
263
+ "gpu": 1.568873804117643e-07,
264
+ "total": 2.9153317899819927e-07
265
  },
266
  "efficiency": {
267
  "unit": "samples/kWh",
268
+ "value": 3430141.308225424
269
  }
270
  }
271
  }