IlyasMoutawwakil HF staff commited on
Commit
089006b
·
verified ·
1 Parent(s): cbf75c8

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -83,7 +83,7 @@
83
  "gpu_count": 1,
84
  "gpu_vram_mb": 68702699520,
85
  "optimum_benchmark_version": "0.2.0",
86
- "optimum_benchmark_commit": "e65976e2695b67f37a76baa42b75347a3a733547",
87
  "transformers_version": "4.40.2",
88
  "transformers_commit": null,
89
  "accelerate_version": "0.30.1",
@@ -102,165 +102,153 @@
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
- "max_ram": 1008.586752,
106
- "max_global_vram": 1877.516288,
107
- "max_process_vram": 232966.889472,
108
  "max_reserved": 555.74528,
109
  "max_allocated": 499.443712
110
  },
111
  "latency": {
112
  "unit": "s",
113
- "count": 137,
114
- "total": 0.9973862705230712,
115
- "mean": 0.0072801917556428565,
116
- "stdev": 0.00023952705644979445,
117
- "p50": 0.007160490989685059,
118
- "p90": 0.00759425106048584,
119
- "p95": 0.007729419231414795,
120
- "p99": 0.008023986015319824,
121
  "values": [
122
- 0.007745292186737061,
123
- 0.007477130889892578,
124
- 0.007263689994812012,
125
- 0.007044650077819824,
126
- 0.007075210094451904,
127
- 0.007119690895080567,
128
- 0.0071025710105896,
129
- 0.007162409782409668,
130
- 0.007113931179046631,
131
- 0.007097609996795654,
132
- 0.007102410793304443,
133
- 0.007073929786682129,
134
- 0.0070678510665893556,
135
- 0.007101930141448974,
136
- 0.007122090816497803,
137
- 0.00716417121887207,
138
- 0.007247690200805664,
139
- 0.007101931095123291,
140
- 0.007119850158691406,
141
- 0.007075049877166748,
142
- 0.007120650768280029,
143
- 0.00708912992477417,
144
- 0.007138570785522461,
145
- 0.007095369815826416,
146
- 0.007122251033782959,
147
- 0.007096010208129883,
148
- 0.007151371002197266,
149
- 0.007112330913543701,
150
- 0.00715680980682373,
151
- 0.008091691970825196,
152
- 0.007667212009429931,
153
- 0.007131370067596435,
154
- 0.0070819311141967775,
155
- 0.007105770111083984,
156
- 0.007234890937805175,
157
- 0.007438411235809326,
158
- 0.007288811206817627,
159
- 0.007059529781341552,
160
- 0.007088490962982178,
161
- 0.0070894498825073244,
162
- 0.007080009937286377,
163
- 0.0073057708740234375,
164
- 0.007294411182403564,
165
- 0.007049449920654297,
166
- 0.007095530986785888,
167
- 0.007088329792022705,
168
- 0.007089611053466797,
169
- 0.0070612897872924804,
170
- 0.00709873104095459,
171
- 0.007068170070648193,
172
- 0.007080330848693848,
173
- 0.007079209804534912,
174
- 0.007053610801696778,
175
- 0.007223851203918457,
176
- 0.007242090225219726,
177
- 0.00704577112197876,
178
- 0.007083690166473389,
179
- 0.007020330905914306,
180
- 0.007039370059967041,
181
- 0.007156811237335205,
182
- 0.007258730888366699,
183
- 0.007137290000915527,
184
- 0.00707905101776123,
185
- 0.007047530174255371,
186
- 0.007334890842437744,
187
- 0.007160490989685059,
188
- 0.0075468912124633785,
189
- 0.007918571949005127,
190
- 0.007079850196838379,
191
- 0.0070592107772827145,
192
- 0.007590091228485108,
193
- 0.007602090835571289,
194
- 0.007431050777435303,
195
- 0.007401771068572998,
196
- 0.007601771831512451,
197
- 0.0075577712059021,
198
- 0.007725450992584228,
199
- 0.0074806509017944335,
200
- 0.007617610931396485,
201
- 0.007434091091156006,
202
- 0.007472970962524414,
203
- 0.007379371166229248,
204
- 0.007331050872802734,
205
- 0.007425450801849366,
206
- 0.007289930820465088,
207
- 0.007331210136413574,
208
- 0.0073494501113891605,
209
- 0.0073828911781311036,
210
- 0.0073737711906433105,
211
- 0.0073449711799621585,
212
- 0.007259050846099854,
213
- 0.007452331066131592,
214
- 0.007371211051940918,
215
- 0.007443050861358643,
216
- 0.007668170928955078,
217
- 0.007532970905303955,
218
- 0.007547211170196533,
219
- 0.007426731109619141,
220
- 0.007456651210784912,
221
- 0.0074880108833312985,
222
- 0.0075939311981201175,
223
- 0.007450251102447509,
224
- 0.007322090148925781,
225
- 0.007592172145843506,
226
- 0.007262570858001709,
227
- 0.007305769920349121,
228
- 0.007168810844421386,
229
- 0.007153131008148193,
230
- 0.007122729778289795,
231
- 0.00717409086227417,
232
- 0.007112810134887695,
233
- 0.007135529994964599,
234
- 0.007139531135559082,
235
- 0.007098090171813965,
236
- 0.007112650871276855,
237
- 0.007113770008087158,
238
- 0.007081770896911621,
239
- 0.007092329978942871,
240
- 0.0070568108558654785,
241
- 0.00710000991821289,
242
- 0.007066411018371582,
243
- 0.007037290096282959,
244
- 0.007108651161193848,
245
- 0.007118889808654785,
246
- 0.0070816102027893065,
247
- 0.007052969932556152,
248
- 0.007987690925598144,
249
- 0.007747372150421142,
250
- 0.007503850936889649,
251
- 0.007542730808258057,
252
- 0.00750321102142334,
253
- 0.007503850936889649,
254
- 0.007594730854034424,
255
- 0.00744209098815918,
256
- 0.007473771095275879,
257
- 0.00802721118927002,
258
- 0.008018252372741699
259
  ]
260
  },
261
  "throughput": {
262
  "unit": "samples/s",
263
- "value": 137.35901931771272
264
  },
265
  "energy": null,
266
  "efficiency": null
 
83
  "gpu_count": 1,
84
  "gpu_vram_mb": 68702699520,
85
  "optimum_benchmark_version": "0.2.0",
86
+ "optimum_benchmark_commit": "519f5ad32b6cd1fea2bd501503e24d19a1e23c22",
87
  "transformers_version": "4.40.2",
88
  "transformers_commit": null,
89
  "accelerate_version": "0.30.1",
 
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 1008.939008,
106
+ "max_global_vram": 2522.841088,
107
+ "max_process_vram": 228416.258048,
108
  "max_reserved": 555.74528,
109
  "max_allocated": 499.443712
110
  },
111
  "latency": {
112
  "unit": "s",
113
+ "count": 125,
114
+ "total": 1.0007468638420107,
115
+ "mean": 0.008005974910736085,
116
+ "stdev": 0.000399916502688993,
117
+ "p50": 0.007859181880950928,
118
+ "p90": 0.008531468200683595,
119
+ "p95": 0.008612684440612793,
120
+ "p99": 0.009603883018493653,
121
  "values": [
122
+ 0.007889421939849853,
123
+ 0.008001102447509765,
124
+ 0.008170061111450196,
125
+ 0.00853774070739746,
126
+ 0.008614700317382812,
127
+ 0.008547981262207032,
128
+ 0.008542059898376465,
129
+ 0.008534860610961915,
130
+ 0.00857374095916748,
131
+ 0.008317900657653808,
132
+ 0.008277741432189942,
133
+ 0.008170541763305663,
134
+ 0.008153902053833009,
135
+ 0.008114702224731445,
136
+ 0.008046701431274414,
137
+ 0.007868782997131347,
138
+ 0.007795661926269531,
139
+ 0.00776734209060669,
140
+ 0.007797102928161621,
141
+ 0.007974222183227539,
142
+ 0.00804094123840332,
143
+ 0.008040461540222168,
144
+ 0.007956942081451417,
145
+ 0.007961902141571045,
146
+ 0.007924302101135254,
147
+ 0.008131341934204101,
148
+ 0.008526379585266114,
149
+ 0.008604620933532715,
150
+ 0.00851006031036377,
151
+ 0.00967069911956787,
152
+ 0.008749900817871093,
153
+ 0.008204940795898438,
154
+ 0.007946221828460694,
155
+ 0.007849102020263671,
156
+ 0.007853423118591309,
157
+ 0.007850702762603759,
158
+ 0.007835182189941406,
159
+ 0.007799662113189697,
160
+ 0.007787502765655517,
161
+ 0.007793581962585449,
162
+ 0.007781261920928955,
163
+ 0.007763503074645996,
164
+ 0.007845582962036133,
165
+ 0.007767501831054687,
166
+ 0.007817262172698975,
167
+ 0.008623661041259766,
168
+ 0.007914381980895996,
169
+ 0.007747982025146484,
170
+ 0.00781630277633667,
171
+ 0.00783822202682495,
172
+ 0.007842542171478272,
173
+ 0.007886062145233155,
174
+ 0.007862382888793946,
175
+ 0.00784030294418335,
176
+ 0.00785134220123291,
177
+ 0.007828621864318848,
178
+ 0.00780974292755127,
179
+ 0.007851663112640381,
180
+ 0.007842542171478272,
181
+ 0.007863982200622559,
182
+ 0.007843982219696045,
183
+ 0.007899342060089111,
184
+ 0.007864142894744873,
185
+ 0.007883662223815919,
186
+ 0.007806702136993408,
187
+ 0.007805262088775635,
188
+ 0.007846063137054443,
189
+ 0.008780940055847168,
190
+ 0.00785790205001831,
191
+ 0.007838702201843262,
192
+ 0.007859181880950928,
193
+ 0.00788670301437378,
194
+ 0.007883022785186768,
195
+ 0.007848941802978515,
196
+ 0.007809902191162109,
197
+ 0.007853422164916992,
198
+ 0.007864462852478027,
199
+ 0.007839023113250732,
200
+ 0.007857741832733155,
201
+ 0.007788462162017822,
202
+ 0.007895823001861573,
203
+ 0.00784350299835205,
204
+ 0.007850542068481445,
205
+ 0.007860941886901856,
206
+ 0.007854221820831299,
207
+ 0.007796782970428467,
208
+ 0.007828941822052002,
209
+ 0.007821422100067138,
210
+ 0.007844781875610352,
211
+ 0.007820302963256837,
212
+ 0.007835981845855712,
213
+ 0.007835502147674561,
214
+ 0.007860621929168702,
215
+ 0.007817263126373292,
216
+ 0.007866062164306641,
217
+ 0.007836462020874024,
218
+ 0.007842862129211425,
219
+ 0.007844462871551513,
220
+ 0.007871502876281739,
221
+ 0.00781534194946289,
222
+ 0.007842862129211425,
223
+ 0.007856942176818848,
224
+ 0.007843983173370362,
225
+ 0.007862862110137939,
226
+ 0.0078113431930541995,
227
+ 0.007892141819000244,
228
+ 0.007884302139282227,
229
+ 0.009392298698425294,
230
+ 0.008292620658874511,
231
+ 0.010851975440979003,
232
+ 0.008164141654968262,
233
+ 0.007853422164916992,
234
+ 0.007854062080383301,
235
+ 0.007792462825775146,
236
+ 0.007847822189331055,
237
+ 0.00788030195236206,
238
+ 0.007861581802368164,
239
+ 0.007910702228546142,
240
+ 0.007918702125549317,
241
+ 0.007863022804260255,
242
+ 0.007855502128601075,
243
+ 0.007899662017822266,
244
+ 0.007852622032165527,
245
+ 0.007864622116088868,
246
+ 0.00790702199935913
 
 
 
 
 
 
 
 
 
 
 
 
247
  ]
248
  },
249
  "throughput": {
250
  "unit": "samples/s",
251
+ "value": 124.90671169341175
252
  },
253
  "energy": null,
254
  "efficiency": null