IlyasMoutawwakil HF staff commited on
Commit
0cbf7f8
1 Parent(s): cc36f49

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -80,7 +80,7 @@
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
- "optimum_benchmark_commit": "43734676b09236e2ae4d7c19f90e8e23d1f2f201",
84
  "transformers_version": "4.42.3",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.31.0",
@@ -99,7 +99,7 @@
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 937.975808,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -107,61 +107,59 @@
107
  },
108
  "latency": {
109
  "unit": "s",
110
- "count": 30,
111
- "total": 1.0257676370003423,
112
- "mean": 0.03419225456667808,
113
- "stdev": 0.0023976719107758203,
114
- "p50": 0.035148123000027454,
115
- "p90": 0.036672888499998633,
116
- "p95": 0.03719636250003475,
117
- "p99": 0.03762319837001428,
118
  "values": [
119
- 0.03392957599999136,
120
- 0.03351084500002344,
121
- 0.03274342299999944,
122
- 0.03230180999997856,
123
- 0.03250899700003629,
124
- 0.03504266100003406,
125
- 0.0377717210000128,
126
- 0.035738279000042894,
127
- 0.0357716799999821,
128
- 0.03711910400005536,
129
- 0.03610407200000054,
130
- 0.03633749599998737,
131
- 0.03662330899999233,
132
- 0.03646065600003112,
133
- 0.03617478200004598,
134
- 0.03613155199997209,
135
- 0.036097679000022254,
136
- 0.036150367000004735,
137
- 0.03637638899999729,
138
- 0.03525358500002085,
139
- 0.03725957400001789,
140
- 0.03260309199998801,
141
- 0.03114338100004943,
142
- 0.03065653199996632,
143
- 0.031438390000005256,
144
- 0.030293444999983876,
145
- 0.030896329000029255,
146
- 0.030805740000005244,
147
- 0.03141648900003702,
148
- 0.031106682000029195
149
  ]
150
  },
151
  "throughput": {
152
  "unit": "samples/s",
153
- "value": 29.246389647980273
154
  },
155
  "energy": {
156
  "unit": "kWh",
157
- "cpu": 1.1996412518048528e-06,
158
- "ram": 5.0135205342436476e-08,
159
  "gpu": 0.0,
160
- "total": 1.2497764571472892e-06
161
  },
162
  "efficiency": {
163
  "unit": "samples/kWh",
164
- "value": 800143.0930156716
165
  }
166
  }
167
  }
 
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
+ "optimum_benchmark_commit": "074fc31a7e13e70facbebbf8b363a0c312f75f69",
84
  "transformers_version": "4.42.3",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.31.0",
 
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 938.758144,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
107
  },
108
  "latency": {
109
  "unit": "s",
110
+ "count": 28,
111
+ "total": 1.0274209120001387,
112
+ "mean": 0.03669360400000495,
113
+ "stdev": 0.00293857696974989,
114
+ "p50": 0.036541684000013674,
115
+ "p90": 0.039925943700006884,
116
+ "p95": 0.040303833900000544,
117
+ "p99": 0.043509746230035375,
118
  "values": [
119
+ 0.03678366599996252,
120
+ 0.03596765200001073,
121
+ 0.036595133999981044,
122
+ 0.034793750999995154,
123
+ 0.03710849200001576,
124
+ 0.03668245599999409,
125
+ 0.03619107000002941,
126
+ 0.0364846080000234,
127
+ 0.036073860999977114,
128
+ 0.03636741899998697,
129
+ 0.040396551000014824,
130
+ 0.04466120200004298,
131
+ 0.03887445899999875,
132
+ 0.03788784700003589,
133
+ 0.03983778600002097,
134
+ 0.03862789800001565,
135
+ 0.039473497999949814,
136
+ 0.040131644999974014,
137
+ 0.03818202600001541,
138
+ 0.03537618799998654,
139
+ 0.036480930000038825,
140
+ 0.03725657900002943,
141
+ 0.036488234000046305,
142
+ 0.03516670599998406,
143
+ 0.03145521900000858,
144
+ 0.03151321799998641,
145
+ 0.030745813999999427,
146
+ 0.031817003000014665
 
 
147
  ]
148
  },
149
  "throughput": {
150
  "unit": "samples/s",
151
+ "value": 27.252705948422648
152
  },
153
  "energy": {
154
  "unit": "kWh",
155
+ "cpu": 1.1956285145949033e-06,
156
+ "ram": 4.9967343072128786e-08,
157
  "gpu": 0.0,
158
+ "total": 1.2455958576670322e-06
159
  },
160
  "efficiency": {
161
  "unit": "samples/kWh",
162
+ "value": 802828.6171993003
163
  }
164
  }
165
  }