IlyasMoutawwakil HF staff commited on
Commit
364a8d7
·
verified ·
1 Parent(s): ca542c3

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -83,7 +83,7 @@
83
  "gpu_count": 1,
84
  "gpu_vram_mb": 68702699520,
85
  "optimum_benchmark_version": "0.2.0",
86
- "optimum_benchmark_commit": "e65976e2695b67f37a76baa42b75347a3a733547",
87
  "transformers_version": "4.40.2",
88
  "transformers_commit": null,
89
  "accelerate_version": "0.30.1",
@@ -102,101 +102,104 @@
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
- "max_ram": 1027.059712,
106
- "max_global_vram": 2103.218176,
107
- "max_process_vram": 287487.868928,
108
  "max_reserved": 773.849088,
109
  "max_allocated": 745.087488
110
  },
111
  "latency": {
112
  "unit": "s",
113
- "count": 73,
114
- "total": 1.0037595462799072,
115
- "mean": 0.013750130770957632,
116
- "stdev": 0.0008642401920468299,
117
- "p50": 0.013329300880432129,
118
- "p90": 0.014802584075927733,
119
- "p95": 0.015227544021606443,
120
- "p99": 0.01577546241760254,
121
  "values": [
122
- 0.014092662811279296,
123
- 0.015107383728027344,
124
- 0.015328024864196777,
125
- 0.015354744911193848,
126
- 0.015321624755859376,
127
- 0.015115864753723144,
128
- 0.01480290412902832,
129
- 0.01475538444519043,
130
- 0.014742263793945313,
131
- 0.014716183662414551,
132
- 0.014111223220825196,
133
- 0.01422722339630127,
134
- 0.013916023254394531,
135
- 0.013079060554504394,
136
- 0.016857307434082033,
137
- 0.013413142204284668,
138
- 0.013090741157531739,
139
- 0.012848660469055176,
140
- 0.013131702423095703,
141
- 0.012845940589904785,
142
- 0.013139861106872559,
143
- 0.012966740608215331,
144
- 0.013175381660461425,
145
- 0.012872980117797852,
146
- 0.013147380828857422,
147
- 0.012934901237487794,
148
- 0.013261621475219726,
149
- 0.012984662055969239,
150
- 0.01326914119720459,
151
- 0.0131299409866333,
152
- 0.013329300880432129,
153
- 0.012940820693969726,
154
- 0.012816981315612793,
155
- 0.013320181846618652,
156
- 0.012951061248779297,
157
- 0.012953141212463378,
158
- 0.01339730167388916,
159
- 0.013113620758056641,
160
- 0.012929460525512696,
161
- 0.013280661582946777,
162
- 0.013672982215881348,
163
- 0.013354262351989747,
164
- 0.013537781715393066,
165
- 0.012857460975646972,
166
- 0.012926740646362305,
167
- 0.013293460845947265,
168
- 0.01286738109588623,
169
- 0.012879220962524415,
170
- 0.013293461799621582,
171
- 0.012923380851745606,
172
- 0.012892661094665527,
173
- 0.013275860786437988,
174
- 0.012925620079040528,
175
- 0.013034741401672363,
176
- 0.01301794147491455,
177
- 0.012966740608215331,
178
- 0.014103062629699707,
179
- 0.014473302841186524,
180
- 0.014006743431091309,
181
- 0.013981463432312012,
182
- 0.014477622985839845,
183
- 0.015164823532104492,
184
- 0.013976503372192382,
185
- 0.014581144332885742,
186
- 0.014375383377075196,
187
- 0.014503543853759765,
188
- 0.014476983070373536,
189
- 0.014390583992004394,
190
- 0.014749624252319337,
191
- 0.01443202304840088,
192
- 0.014412342071533203,
193
- 0.014360183715820313,
194
- 0.01480130386352539
 
 
 
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
- "value": 72.72658105275276
200
  },
201
  "energy": null,
202
  "efficiency": null
 
83
  "gpu_count": 1,
84
  "gpu_vram_mb": 68702699520,
85
  "optimum_benchmark_version": "0.2.0",
86
+ "optimum_benchmark_commit": "519f5ad32b6cd1fea2bd501503e24d19a1e23c22",
87
  "transformers_version": "4.40.2",
88
  "transformers_commit": null,
89
  "accelerate_version": "0.30.1",
 
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 1026.02752,
106
+ "max_global_vram": 2748.563456,
107
+ "max_process_vram": 291436.978176,
108
  "max_reserved": 773.849088,
109
  "max_allocated": 745.087488
110
  },
111
  "latency": {
112
  "unit": "s",
113
+ "count": 76,
114
+ "total": 1.007524724006653,
115
+ "mean": 0.01325690426324543,
116
+ "stdev": 0.0006684026301356472,
117
+ "p50": 0.012951571941375733,
118
+ "p90": 0.014346209526062012,
119
+ "p95": 0.014461408853530883,
120
+ "p99": 0.01484936785697937,
121
  "values": [
122
+ 0.015529887199401855,
123
+ 0.014364768981933593,
124
+ 0.014325408935546876,
125
+ 0.014318689346313477,
126
+ 0.014326370239257813,
127
+ 0.01433388900756836,
128
+ 0.014277249336242677,
129
+ 0.014437249183654785,
130
+ 0.014348770141601562,
131
+ 0.014451008796691894,
132
+ 0.014317568778991699,
133
+ 0.014492609024047852,
134
+ 0.014343648910522462,
135
+ 0.01308701229095459,
136
+ 0.012704453468322753,
137
+ 0.012703012466430665,
138
+ 0.012919013023376464,
139
+ 0.012730691909790039,
140
+ 0.012754053115844726,
141
+ 0.012974692344665526,
142
+ 0.012708292961120606,
143
+ 0.012783013343811035,
144
+ 0.012948612213134765,
145
+ 0.012770691871643066,
146
+ 0.012777732849121095,
147
+ 0.012994372367858886,
148
+ 0.012730692863464356,
149
+ 0.012721412658691407,
150
+ 0.012996453285217285,
151
+ 0.012794692039489746,
152
+ 0.012742853164672852,
153
+ 0.012756612777709961,
154
+ 0.012696453094482422,
155
+ 0.01301501178741455,
156
+ 0.012778372764587402,
157
+ 0.01276941204071045,
158
+ 0.01300285243988037,
159
+ 0.01273757266998291,
160
+ 0.012778853416442871,
161
+ 0.012975491523742676,
162
+ 0.012681093215942383,
163
+ 0.01273869228363037,
164
+ 0.013248612403869629,
165
+ 0.012954531669616699,
166
+ 0.012799972534179688,
167
+ 0.013232770919799805,
168
+ 0.012769573211669922,
169
+ 0.012782692909240723,
170
+ 0.013044452667236328,
171
+ 0.012795493125915528,
172
+ 0.012836612701416016,
173
+ 0.013113571166992187,
174
+ 0.012787013053894044,
175
+ 0.012770052909851073,
176
+ 0.014622528076171875,
177
+ 0.01336925220489502,
178
+ 0.013881409645080566,
179
+ 0.01456044864654541,
180
+ 0.01358428955078125,
181
+ 0.013572291374206543,
182
+ 0.013431971549987792,
183
+ 0.013588770866394043,
184
+ 0.013832450866699219,
185
+ 0.013484451293945312,
186
+ 0.012767011642456055,
187
+ 0.012796612739562988,
188
+ 0.013047492027282715,
189
+ 0.012832932472229004,
190
+ 0.013159972190856933,
191
+ 0.012858053207397462,
192
+ 0.012831011772155761,
193
+ 0.012857253074645997,
194
+ 0.013220610618591309,
195
+ 0.01284013271331787,
196
+ 0.01280989170074463,
197
+ 0.012801252365112304
198
  ]
199
  },
200
  "throughput": {
201
  "unit": "samples/s",
202
+ "value": 75.43239206852274
203
  },
204
  "energy": null,
205
  "efficiency": null