IlyasMoutawwakil HF staff commited on
Commit
da6c12f
1 Parent(s): 60e3c6f

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -78,7 +78,7 @@
78
  "processor": "x86_64",
79
  "python_version": "3.10.14",
80
  "optimum_benchmark_version": "0.2.0",
81
- "optimum_benchmark_commit": "c78698d47f8aff03fc8162bfacd2946f5882e025",
82
  "transformers_version": "4.40.2",
83
  "transformers_commit": null,
84
  "accelerate_version": "0.30.1",
@@ -97,7 +97,7 @@
97
  "forward": {
98
  "memory": {
99
  "unit": "MB",
100
- "max_ram": 936.767488,
101
  "max_global_vram": null,
102
  "max_process_vram": null,
103
  "max_reserved": null,
@@ -105,59 +105,60 @@
105
  },
106
  "latency": {
107
  "unit": "s",
108
- "count": 28,
109
- "total": 1.0101344830001722,
110
- "mean": 0.036076231535720434,
111
- "stdev": 0.000694139201075085,
112
- "p50": 0.03610273799998254,
113
- "p90": 0.0368631685000139,
114
- "p95": 0.0370095093000117,
115
- "p99": 0.03728659555002707,
116
  "values": [
117
- 0.036092067999959454,
118
- 0.03657427499996402,
119
- 0.03627225300004966,
120
- 0.037370143000032385,
121
- 0.03492255499998009,
122
- 0.036914426000009826,
123
- 0.036810113000001365,
124
- 0.03633397800001603,
125
- 0.036814120000030925,
126
- 0.03706070800001271,
127
- 0.035931369000024915,
128
- 0.03616614599997092,
129
- 0.036841201000015644,
130
- 0.035238772000013796,
131
- 0.035587830999986636,
132
- 0.035485891999996966,
133
- 0.03598002900002939,
134
- 0.035328329000037684,
135
- 0.03523166899998387,
136
- 0.03447857999998405,
137
- 0.03579876300000251,
138
- 0.03569773500004203,
139
- 0.03595847999997659,
140
- 0.03611340800000562,
141
- 0.035329200000035144,
142
- 0.0367212780000159,
143
- 0.0365498379999849,
144
- 0.036531324000009135
 
145
  ]
146
  },
147
  "throughput": {
148
  "unit": "samples/s",
149
- "value": 27.719081440362263
150
  },
151
  "energy": {
152
  "unit": "kWh",
153
- "cpu": 1.2407072302367953e-06,
154
- "ram": 5.18503560006689e-08,
155
  "gpu": 0.0,
156
- "total": 1.2925575862374642e-06
157
  },
158
  "efficiency": {
159
  "unit": "samples/kWh",
160
- "value": 773659.9209563445
161
  }
162
  }
163
  }
 
78
  "processor": "x86_64",
79
  "python_version": "3.10.14",
80
  "optimum_benchmark_version": "0.2.0",
81
+ "optimum_benchmark_commit": "83a87c3d95805880a695f5bd4e2d038143845010",
82
  "transformers_version": "4.40.2",
83
  "transformers_commit": null,
84
  "accelerate_version": "0.30.1",
 
97
  "forward": {
98
  "memory": {
99
  "unit": "MB",
100
+ "max_ram": 937.234432,
101
  "max_global_vram": null,
102
  "max_process_vram": null,
103
  "max_reserved": null,
 
105
  },
106
  "latency": {
107
  "unit": "s",
108
+ "count": 29,
109
+ "total": 1.0045533799998907,
110
+ "mean": 0.03463977172413416,
111
+ "stdev": 0.001843224769988702,
112
+ "p50": 0.03498452700000598,
113
+ "p90": 0.036115662200006685,
114
+ "p95": 0.03655359919998773,
115
+ "p99": 0.036731589599990004,
116
  "values": [
117
+ 0.03563756799997009,
118
+ 0.03554033599999684,
119
+ 0.03572528099999772,
120
+ 0.03482473899998695,
121
+ 0.0352656320000051,
122
+ 0.034991369000010764,
123
+ 0.03519440900004156,
124
+ 0.034824958000001516,
125
+ 0.034968736999985595,
126
+ 0.03470389200003865,
127
+ 0.03474972800000842,
128
+ 0.035279397000010704,
129
+ 0.03498452700000598,
130
+ 0.03484873299998981,
131
+ 0.03491746100002047,
132
+ 0.034666372000003776,
133
+ 0.0364454869999804,
134
+ 0.03569689900001549,
135
+ 0.035510088999956224,
136
+ 0.03603320600001325,
137
+ 0.03486565499997596,
138
+ 0.03581586999996489,
139
+ 0.036625673999992614,
140
+ 0.036772778999988986,
141
+ 0.034601658999974916,
142
+ 0.030703904999995757,
143
+ 0.030261446999986674,
144
+ 0.02988289000001032,
145
+ 0.030214680999961274
146
  ]
147
  },
148
  "throughput": {
149
  "unit": "samples/s",
150
+ "value": 28.86855051943895
151
  },
152
  "energy": {
153
  "unit": "kWh",
154
+ "cpu": 1.1788723444697834e-06,
155
+ "ram": 4.926668488904494e-08,
156
  "gpu": 0.0,
157
+ "total": 1.2281390293588283e-06
158
  },
159
  "efficiency": {
160
  "unit": "samples/kWh",
161
+ "value": 814240.0624806035
162
  }
163
  }
164
  }