IlyasMoutawwakil HF staff commited on
Commit
2c01266
·
verified ·
1 Parent(s): 464d80d

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -85,7 +85,7 @@
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.2.1",
88
- "optimum_benchmark_commit": "9a8b29987613b8b04c221447a49b37ee314548ff",
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.31.0",
@@ -104,156 +104,168 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1010.040832,
108
- "max_global_vram": 897.687552,
109
- "max_process_vram": 200340.6848,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.443712
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 128,
116
- "total": 0.9970108065605162,
117
- "mean": 0.007789146926254034,
118
- "stdev": 0.00034209279895787794,
119
- "p50": 0.007701303958892822,
120
- "p90": 0.007846343851089477,
121
- "p95": 0.00826522626876831,
122
- "p99": 0.009414378137588502,
123
  "values": [
124
- 0.00775874376296997,
125
- 0.007798744201660156,
126
- 0.007833623886108398,
127
- 0.007767545223236084,
128
- 0.007738264083862305,
129
- 0.007688024044036865,
130
- 0.007708663940429687,
131
- 0.0076896238327026365,
132
- 0.007742424011230469,
133
- 0.007776185035705567,
134
- 0.007721144199371338,
135
- 0.007721784114837646,
136
- 0.007790744781494141,
137
- 0.007808184146881104,
138
- 0.007761464118957519,
139
- 0.009003707885742188,
140
- 0.007906905174255371,
141
- 0.007713944911956787,
142
- 0.007692024230957031,
143
- 0.007695703983306885,
144
- 0.007681784152984619,
145
- 0.007666903972625733,
146
- 0.007628663063049317,
147
- 0.0076438641548156735,
148
- 0.007644023895263672,
149
- 0.008144024848937989,
150
- 0.008569626808166505,
151
- 0.0077225852012634275,
152
- 0.007819704055786133,
153
- 0.007643383979797363,
154
- 0.008008984565734864,
155
- 0.007692344188690185,
156
- 0.007642903804779053,
157
- 0.007633463859558105,
158
- 0.007668504238128662,
159
- 0.007625944137573242,
160
- 0.007636983871459961,
161
- 0.0076272239685058595,
162
- 0.007615863800048828,
163
- 0.007668505191802979,
164
- 0.007675704002380371,
165
- 0.007681303977966308,
166
- 0.007664185047149658,
167
- 0.007655064105987549,
168
- 0.007686264038085937,
169
- 0.00765442419052124,
170
- 0.0076469039916992184,
171
- 0.007645143985748291,
172
- 0.00761954402923584,
173
- 0.007624184131622315,
174
- 0.007676504135131836,
175
- 0.007653304100036621,
176
- 0.007643383979797363,
177
- 0.0076560239791870115,
178
- 0.008727867126464844,
179
- 0.008203866004943848,
180
- 0.009566269874572753,
181
- 0.007896504878997802,
182
- 0.007733304977416992,
183
- 0.007708183765411377,
184
- 0.007693943977355957,
185
- 0.007669784069061279,
186
- 0.0076584239006042484,
187
- 0.007672023773193359,
188
- 0.00771938419342041,
189
- 0.007654263973236084,
190
- 0.0076584239006042484,
191
- 0.007671703815460205,
192
- 0.007678904056549072,
193
- 0.007678103923797607,
194
- 0.007654104232788086,
195
- 0.007683703899383545,
196
- 0.0076534638404846195,
197
- 0.007734904766082764,
198
- 0.0076859440803527836,
199
- 0.007664504051208496,
200
- 0.007677624225616455,
201
- 0.007667543888092041,
202
- 0.007737464904785156,
203
- 0.007703224182128906,
204
- 0.007669623851776123,
205
- 0.007725625038146972,
206
- 0.007678423881530762,
207
- 0.007741943836212158,
208
- 0.0077843451499938964,
209
- 0.007708183765411377,
210
- 0.007697944164276123,
211
- 0.0077241849899292,
212
- 0.007764503955841064,
213
- 0.007744665145874023,
214
- 0.007759384155273438,
215
- 0.0077361841201782225,
216
- 0.007734903812408447,
217
- 0.007728663921356202,
218
- 0.007794105052947998,
219
- 0.00891554832458496,
220
- 0.008298266410827636,
221
- 0.007813304901123046,
222
- 0.007780824184417725,
223
- 0.007766744136810303,
224
- 0.007741464138031006,
225
- 0.007734904766082764,
226
- 0.007742424011230469,
227
- 0.0077195439338684085,
228
- 0.007701623916625977,
229
- 0.0077241849899292,
230
- 0.007703224182128906,
231
- 0.007699063777923584,
232
- 0.007681943893432617,
233
- 0.007673624038696289,
234
- 0.007730743885040283,
235
- 0.0077459440231323244,
236
- 0.007737143993377686,
237
- 0.0076982641220092775,
238
- 0.0076889839172363285,
239
- 0.007648983955383301,
240
- 0.007693143844604492,
241
- 0.007692344188690185,
242
- 0.007685463905334472,
243
- 0.0076792240142822266,
244
- 0.007696023941040039,
245
- 0.01019923210144043,
246
- 0.007876023769378663,
247
- 0.007720344066619873,
248
- 0.007700984001159668,
249
- 0.007726425170898437,
250
- 0.0077221040725708005,
251
- 0.007741943836212158
 
 
 
 
 
 
 
 
 
 
 
 
252
  ]
253
  },
254
  "throughput": {
255
  "unit": "samples/s",
256
- "value": 128.38376390480045
257
  },
258
  "energy": null,
259
  "efficiency": null
 
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": "3b8c49a169ebd79001b2a83fbf2b332612417102",
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.31.0",
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 1009.905664,
108
+ "max_global_vram": 897.138688,
109
+ "max_process_vram": 195782.180864,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.443712
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 140,
116
+ "total": 0.9992238569259644,
117
+ "mean": 0.007137313263756888,
118
+ "stdev": 0.0006227442120522962,
119
+ "p50": 0.006814506530761719,
120
+ "p90": 0.007902210140228271,
121
+ "p95": 0.008003529357910155,
122
+ "p99": 0.009309260616302484,
123
  "values": [
124
+ 0.009456087112426757,
125
+ 0.009468565940856934,
126
+ 0.009079608917236328,
127
+ 0.008036577224731445,
128
+ 0.008018497467041016,
129
+ 0.00801401710510254,
130
+ 0.00800297737121582,
131
+ 0.007981698036193847,
132
+ 0.007909698009490967,
133
+ 0.00794233798980713,
134
+ 0.007817859172821046,
135
+ 0.007829699039459228,
136
+ 0.007881857872009278,
137
+ 0.007912417888641358,
138
+ 0.007898978233337402,
139
+ 0.007901378154754638,
140
+ 0.007883617877960205,
141
+ 0.007853377819061279,
142
+ 0.007878818988800048,
143
+ 0.007847939014434814,
144
+ 0.007961698055267334,
145
+ 0.007945858001708984,
146
+ 0.007900419235229493,
147
+ 0.007892738819122315,
148
+ 0.008969849586486816,
149
+ 0.007810978889465332,
150
+ 0.007608420848846435,
151
+ 0.007575621128082275,
152
+ 0.007582981109619141,
153
+ 0.007572260856628418,
154
+ 0.007542820930480957,
155
+ 0.007507781028747559,
156
+ 0.007535941123962402,
157
+ 0.007516100883483887,
158
+ 0.007528901100158691,
159
+ 0.0075025010108947755,
160
+ 0.007505701065063476,
161
+ 0.007517221927642822,
162
+ 0.007579781055450439,
163
+ 0.007515621185302734,
164
+ 0.007526181221008301,
165
+ 0.007542981147766113,
166
+ 0.0075293807983398435,
167
+ 0.007508262157440186,
168
+ 0.007499141216278076,
169
+ 0.0075165810585021975,
170
+ 0.007562981128692627,
171
+ 0.007577381134033203,
172
+ 0.0075380210876464845,
173
+ 0.00751370096206665,
174
+ 0.007522820949554443,
175
+ 0.007531620979309082,
176
+ 0.007387621879577637,
177
+ 0.006944106101989746,
178
+ 0.006728587150573731,
179
+ 0.006620428085327148,
180
+ 0.006630348205566406,
181
+ 0.006658827781677246,
182
+ 0.006628588199615478,
183
+ 0.006608588218688965,
184
+ 0.006614667892456055,
185
+ 0.006641707897186279,
186
+ 0.006640427112579346,
187
+ 0.006638507843017578,
188
+ 0.006674828052520752,
189
+ 0.0066209077835083004,
190
+ 0.0066245880126953125,
191
+ 0.0066245880126953125,
192
+ 0.0066293878555297855,
193
+ 0.006798986911773682,
194
+ 0.006617708206176758,
195
+ 0.00658810806274414,
196
+ 0.006622347831726074,
197
+ 0.006616108894348145,
198
+ 0.006623467922210693,
199
+ 0.006624427795410157,
200
+ 0.00663402795791626,
201
+ 0.006848266124725342,
202
+ 0.006632748126983642,
203
+ 0.0066338682174682615,
204
+ 0.006629228115081787,
205
+ 0.0066362681388854984,
206
+ 0.0066276278495788574,
207
+ 0.006623467922210693,
208
+ 0.00662746810913086,
209
+ 0.006652586936950684,
210
+ 0.006613708019256591,
211
+ 0.00663354778289795,
212
+ 0.006651308059692383,
213
+ 0.007187304019927979,
214
+ 0.0070804238319396974,
215
+ 0.007057384014129638,
216
+ 0.00704570484161377,
217
+ 0.007540900230407715,
218
+ 0.007566020965576172,
219
+ 0.00733642292022705,
220
+ 0.007087463855743408,
221
+ 0.007017704963684082,
222
+ 0.006962825775146484,
223
+ 0.007042665004730225,
224
+ 0.007106823921203613,
225
+ 0.007033384799957275,
226
+ 0.006884585857391357,
227
+ 0.006618507862091064,
228
+ 0.006651947975158692,
229
+ 0.006630988121032715,
230
+ 0.006599947929382324,
231
+ 0.006606667995452881,
232
+ 0.0065943489074707035,
233
+ 0.006590349197387695,
234
+ 0.006625547885894776,
235
+ 0.006625708103179932,
236
+ 0.0066228280067443844,
237
+ 0.006606028079986572,
238
+ 0.0066221880912780764,
239
+ 0.006639947891235351,
240
+ 0.006649548053741455,
241
+ 0.00663354778289795,
242
+ 0.006642347812652588,
243
+ 0.006626987934112549,
244
+ 0.006626187801361084,
245
+ 0.006605709075927734,
246
+ 0.006613708972930908,
247
+ 0.006677867889404297,
248
+ 0.0068300261497497555,
249
+ 0.0066172280311584475,
250
+ 0.006626028060913086,
251
+ 0.006602828979492188,
252
+ 0.006608908176422119,
253
+ 0.006606828212738037,
254
+ 0.006601707935333252,
255
+ 0.006641228199005127,
256
+ 0.006699306964874268,
257
+ 0.0066551480293273926,
258
+ 0.006640428066253662,
259
+ 0.006641707897186279,
260
+ 0.006636427879333496,
261
+ 0.006613868236541748,
262
+ 0.006625867843627929,
263
+ 0.006605547904968262
264
  ]
265
  },
266
  "throughput": {
267
  "unit": "samples/s",
268
+ "value": 140.10874443160242
269
  },
270
  "energy": null,
271
  "efficiency": null