IlyasMoutawwakil HF staff commited on
Commit
d4fffdb
·
verified ·
1 Parent(s): 904ddc6

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -102,7 +102,7 @@
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
- "max_ram": 975.716352,
106
  "max_global_vram": 1434.976256,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 794.820608,
@@ -110,100 +110,103 @@
110
  },
111
  "latency": {
112
  "unit": "s",
113
- "count": 69,
114
- "total": 1.0024325485229493,
115
- "mean": 0.014528007949607957,
116
- "stdev": 0.000540985829057184,
117
- "p50": 0.014315520286560059,
118
- "p90": 0.015065087890625,
119
- "p95": 0.015385600090026855,
120
- "p99": 0.016900639038085932,
121
  "values": [
122
- 0.015221759796142579,
123
- 0.015686655998229982,
124
- 0.015385600090026855,
125
- 0.015089664459228515,
126
- 0.014782464027404785,
127
- 0.014781439781188965,
128
- 0.014626815795898437,
129
- 0.014215167999267577,
130
- 0.014289919853210448,
131
- 0.01447219181060791,
132
- 0.014315520286560059,
133
- 0.01472000026702881,
134
- 0.015058943748474121,
135
- 0.014732288360595704,
136
- 0.014286848068237304,
137
- 0.01440665626525879,
138
- 0.014233599662780762,
139
- 0.014223360061645507,
140
- 0.014193663597106934,
141
- 0.014227456092834472,
142
- 0.014424063682556153,
143
- 0.014584832191467285,
144
- 0.014435327529907227,
145
- 0.014329888343811036,
146
- 0.014387200355529785,
147
- 0.014359552383422852,
148
- 0.014180352210998535,
149
- 0.01424176025390625,
150
- 0.014509056091308594,
151
- 0.014167039871215821,
152
- 0.014256128311157227,
153
- 0.014140416145324707,
154
- 0.014282719612121582,
155
- 0.01475887966156006,
156
- 0.014569472312927247,
157
- 0.01430016040802002,
158
- 0.014320639610290528,
159
- 0.014315520286560059,
160
- 0.01415782356262207,
161
- 0.014213120460510254,
162
- 0.014021632194519042,
163
- 0.014539775848388671,
164
- 0.017256479263305664,
165
- 0.016733184814453125,
166
- 0.01490124797821045,
167
- 0.01469536018371582,
168
- 0.015385600090026855,
169
- 0.014763008117675782,
170
- 0.014309375762939454,
171
- 0.014323712348937988,
172
- 0.01427455997467041,
173
- 0.01425920009613037,
174
- 0.014493760108947753,
175
- 0.014395392417907715,
176
- 0.014298111915588378,
177
- 0.014243840217590332,
178
- 0.014738431930541992,
179
- 0.014309375762939454,
180
- 0.014224384307861328,
181
- 0.014302240371704102,
182
- 0.014315520286560059,
183
- 0.014311424255371094,
184
- 0.014260224342346191,
185
- 0.01428275203704834,
186
- 0.014313471794128419,
187
- 0.014271488189697265,
188
- 0.014098464012145996,
189
- 0.0140697603225708,
190
- 0.014156800270080566
 
 
 
191
  ]
192
  },
193
  "throughput": {
194
  "unit": "samples/s",
195
- "value": 68.83256145430353
196
  },
197
  "energy": {
198
  "unit": "kWh",
199
- "cpu": 1.6294535890198917e-07,
200
- "ram": 8.90831644838191e-08,
201
- "gpu": 3.410048390136998e-07,
202
- "total": 5.93033362399508e-07
203
  },
204
  "efficiency": {
205
  "unit": "samples/kWh",
206
- "value": 1686245.7719981212
207
  }
208
  }
209
  }
 
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 975.527936,
106
  "max_global_vram": 1434.976256,
107
  "max_process_vram": 0.0,
108
  "max_reserved": 794.820608,
 
110
  },
111
  "latency": {
112
  "unit": "s",
113
+ "count": 72,
114
+ "total": 0.998003776550293,
115
+ "mean": 0.013861163563198513,
116
+ "stdev": 0.0006246735784352315,
117
+ "p50": 0.013868031978607177,
118
+ "p90": 0.014777651405334474,
119
+ "p95": 0.015252582693099976,
120
+ "p99": 0.015793869161605835,
121
  "values": [
122
+ 0.015244288444519043,
123
+ 0.01603379249572754,
124
+ 0.01569587230682373,
125
+ 0.015694848060607912,
126
+ 0.015262720108032226,
127
+ 0.015213567733764649,
128
+ 0.014815232276916505,
129
+ 0.015013855934143066,
130
+ 0.013454336166381836,
131
+ 0.013370368003845215,
132
+ 0.013210783958435059,
133
+ 0.013239295959472656,
134
+ 0.013301759719848634,
135
+ 0.013288448333740235,
136
+ 0.013129695892333985,
137
+ 0.013264896392822266,
138
+ 0.013399040222167969,
139
+ 0.013381631851196289,
140
+ 0.013368320465087891,
141
+ 0.013373439788818359,
142
+ 0.013362175941467285,
143
+ 0.013416447639465333,
144
+ 0.01336723232269287,
145
+ 0.013365216255187988,
146
+ 0.013452287673950195,
147
+ 0.01338368034362793,
148
+ 0.013421567916870117,
149
+ 0.013398015975952148,
150
+ 0.013379584312438965,
151
+ 0.013540384292602538,
152
+ 0.013355008125305176,
153
+ 0.013462528228759766,
154
+ 0.013426624298095702,
155
+ 0.013363264083862304,
156
+ 0.01338368034362793,
157
+ 0.013321215629577637,
158
+ 0.013446144104003906,
159
+ 0.013480959892272949,
160
+ 0.013475839614868163,
161
+ 0.01349728012084961,
162
+ 0.013454303741455079,
163
+ 0.013817855834960938,
164
+ 0.014030847549438476,
165
+ 0.013965312004089356,
166
+ 0.014027775764465332,
167
+ 0.014035008430480956,
168
+ 0.014034943580627441,
169
+ 0.014003199577331543,
170
+ 0.014034943580627441,
171
+ 0.014018560409545898,
172
+ 0.014032896041870118,
173
+ 0.014114815711975098,
174
+ 0.013991999626159667,
175
+ 0.014005248069763183,
176
+ 0.01399295997619629,
177
+ 0.014005248069763183,
178
+ 0.014439423561096192,
179
+ 0.013999103546142578,
180
+ 0.013912063598632812,
181
+ 0.013940735816955567,
182
+ 0.013940735816955567,
183
+ 0.013927424430847168,
184
+ 0.013955072402954101,
185
+ 0.01391107177734375,
186
+ 0.013939680099487305,
187
+ 0.013913087844848633,
188
+ 0.013884415626525879,
189
+ 0.013991935729980469,
190
+ 0.013920224189758301,
191
+ 0.014033920288085937,
192
+ 0.01382198429107666,
193
+ 0.013851648330688476
194
  ]
195
  },
196
  "throughput": {
197
  "unit": "samples/s",
198
+ "value": 72.14401557564814
199
  },
200
  "energy": {
201
  "unit": "kWh",
202
+ "cpu": 1.676801854455975e-07,
203
+ "ram": 9.126176298874705e-08,
204
+ "gpu": 3.5294629174648144e-07,
205
+ "total": 6.11888240180826e-07
206
  },
207
  "efficiency": {
208
  "unit": "samples/kWh",
209
+ "value": 1634285.3716300852
210
  }
211
  }
212
  }