IlyasMoutawwakil HF staff commited on
Commit
57ed2fa
·
verified ·
1 Parent(s): e6bc799

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -104,7 +104,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 976.388096,
108
  "max_global_vram": 1434.976256,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 794.820608,
@@ -112,102 +112,103 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 71,
116
- "total": 1.0102445478439332,
117
- "mean": 0.014228796448506099,
118
- "stdev": 0.0009817016998652183,
119
- "p50": 0.013974528312683105,
120
- "p90": 0.014649344444274901,
121
- "p95": 0.015409631729125977,
122
- "p99": 0.018289459228515617,
123
  "values": [
124
- 0.0204083194732666,
125
- 0.01738137626647949,
126
- 0.015316991806030274,
127
- 0.014405632019042968,
128
- 0.01446399974822998,
129
- 0.014279680252075195,
130
- 0.014346240043640136,
131
- 0.014503935813903808,
132
- 0.01429196834564209,
133
- 0.014495743751525878,
134
- 0.01448755168914795,
135
- 0.014315520286560059,
136
- 0.014387200355529785,
137
- 0.014179327964782714,
138
- 0.014231552124023437,
139
- 0.014306303977966308,
140
- 0.014771200180053711,
141
- 0.016741376876831054,
142
- 0.015292415618896485,
143
- 0.014649344444274901,
144
- 0.014338047981262207,
145
- 0.014272512435913086,
146
- 0.014499839782714843,
147
- 0.014235648155212402,
148
- 0.014286848068237304,
149
- 0.014269439697265626,
150
- 0.01407487964630127,
151
- 0.013990912437438965,
152
- 0.013796352386474609,
153
- 0.013757439613342285,
154
- 0.013666303634643554,
155
- 0.013617152214050294,
156
- 0.013691904067993164,
157
- 0.013463552474975587,
158
- 0.013412351608276368,
159
- 0.013974528312683105,
160
- 0.014211071968078613,
161
- 0.014405632019042968,
162
- 0.014118847846984863,
163
- 0.014025728225708007,
164
  0.01406771183013916,
165
- 0.013800448417663574,
166
- 0.013830143928527832,
167
- 0.013783040046691895,
168
- 0.013743103981018067,
169
- 0.01550227165222168,
170
- 0.014024703979492188,
171
- 0.013835264205932616,
172
- 0.013828160285949707,
173
- 0.013748224258422852,
174
- 0.013799424171447755,
175
- 0.013701120376586913,
176
- 0.013790207862854004,
177
- 0.013723648071289063,
178
- 0.013730815887451172,
179
- 0.013716480255126954,
180
- 0.013709312438964843,
181
- 0.013761568069458007,
182
- 0.013737983703613281,
183
- 0.01377791976928711,
184
- 0.013719552040100098,
185
- 0.013799424171447755,
186
- 0.013747200012207032,
187
- 0.013791232109069825,
188
- 0.013752320289611816,
189
- 0.013728768348693847,
190
- 0.013749216079711914,
191
- 0.013730815887451172,
192
- 0.013733887672424316,
193
- 0.01375641632080078,
194
- 0.013759488105773926
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
- "value": 70.28001304389953
200
  },
201
  "energy": {
202
  "unit": "kWh",
203
- "cpu": 1.6352259631933507e-07,
204
- "ram": 8.935887694218328e-08,
205
- "gpu": 3.4232143063013805e-07,
206
- "total": 5.952029038916563e-07
207
  },
208
  "efficiency": {
209
  "unit": "samples/kWh",
210
- "value": 1680099.32992872
211
  }
212
  }
213
  }
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 976.379904,
108
  "max_global_vram": 1434.976256,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 794.820608,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 72,
116
+ "total": 1.0052177286148074,
117
+ "mean": 0.013961357341872323,
118
+ "stdev": 0.0006157268333519445,
119
+ "p50": 0.01400268793106079,
120
+ "p90": 0.014582879734039307,
121
+ "p95": 0.015084646654129029,
122
+ "p99": 0.01597340756416321,
123
  "values": [
124
+ 0.015462400436401368,
125
+ 0.015742976188659667,
126
+ 0.014993408203125,
127
+ 0.01471180820465088,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
128
  0.01406771183013916,
129
+ 0.014029824256896972,
130
+ 0.014005248069763183,
131
+ 0.014100480079650878,
132
+ 0.014140416145324707,
133
+ 0.014047231674194336,
134
+ 0.014227456092834472,
135
+ 0.014223360061645507,
136
+ 0.014008319854736329,
137
+ 0.014172160148620605,
138
+ 0.01398476791381836,
139
+ 0.014082048416137695,
140
+ 0.014127103805541993,
141
+ 0.016537567138671876,
142
+ 0.015196160316467285,
143
+ 0.014438400268554688,
144
+ 0.014234623908996581,
145
+ 0.014147583961486816,
146
+ 0.014041088104248046,
147
+ 0.014234623908996581,
148
+ 0.014007295608520508,
149
+ 0.014000127792358399,
150
+ 0.014005248069763183,
151
+ 0.014269439697265626,
152
+ 0.01438003158569336,
153
+ 0.01406873607635498,
154
+ 0.014204928398132324,
155
+ 0.013980671882629395,
156
+ 0.01439641571044922,
157
+ 0.01498624038696289,
158
+ 0.013807616233825683,
159
+ 0.014208000183105468,
160
+ 0.014355456352233887,
161
+ 0.014586879730224609,
162
+ 0.014546879768371582,
163
+ 0.014071776390075683,
164
+ 0.013727744102478028,
165
+ 0.01363046360015869,
166
+ 0.013625344276428223,
167
+ 0.013645824432373046,
168
+ 0.013363200187683106,
169
+ 0.013363200187683106,
170
+ 0.013352959632873536,
171
+ 0.013350975990295411,
172
+ 0.01338060760498047,
173
+ 0.013385727882385253,
174
+ 0.013385727882385253,
175
+ 0.013362175941467285,
176
+ 0.01344102382659912,
177
+ 0.013415424346923829,
178
+ 0.0134717435836792,
179
+ 0.01338060760498047,
180
+ 0.013406208038330078,
181
+ 0.01344102382659912,
182
+ 0.01337548828125,
183
+ 0.013386752128601074,
184
+ 0.013385727882385253,
185
+ 0.013439999580383302,
186
+ 0.013378560066223144,
187
+ 0.013386752128601074,
188
+ 0.01336729621887207,
189
+ 0.013458432197570801,
190
+ 0.013378560066223144,
191
+ 0.013353983879089355,
192
+ 0.013612031936645508,
193
+ 0.013660160064697266,
194
+ 0.013656064033508301,
195
+ 0.013415424346923829
196
  ]
197
  },
198
  "throughput": {
199
  "unit": "samples/s",
200
+ "value": 71.62627354296288
201
  },
202
  "energy": {
203
  "unit": "kWh",
204
+ "cpu": 1.6156246175279133e-07,
205
+ "ram": 8.825076518058361e-08,
206
+ "gpu": 3.525453270810337e-07,
207
+ "total": 6.023585540144086e-07
208
  },
209
  "efficiency": {
210
  "unit": "samples/kWh",
211
+ "value": 1660140.780496129
212
  }
213
  }
214
  }