IlyasMoutawwakil HF staff commited on
Commit
e321bcb
1 Parent(s): 5e08020

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -80,7 +80,7 @@
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.0",
83
- "optimum_benchmark_commit": "2d3261cf9b992810a685ce736f41254a91cd091e",
84
  "transformers_version": "4.40.2",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.30.1",
@@ -99,7 +99,7 @@
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 938.47552,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -107,58 +107,62 @@
107
  },
108
  "latency": {
109
  "unit": "s",
110
- "count": 27,
111
- "total": 1.0055812380001043,
112
- "mean": 0.03724374955555942,
113
- "stdev": 0.001135183103969783,
114
- "p50": 0.037510698000005505,
115
- "p90": 0.038273721600000955,
116
- "p95": 0.03854815469999835,
117
- "p99": 0.03917231396003558,
118
  "values": [
119
- 0.03690168700001095,
120
- 0.03562123800003292,
121
- 0.03564668699999629,
122
- 0.03647698100002117,
123
- 0.03408332599997266,
124
- 0.037355728000022737,
125
- 0.037032312000008005,
126
- 0.03689864199998283,
127
- 0.037305964999973185,
128
- 0.037539271000014196,
129
- 0.035390135999989525,
130
- 0.03557062400000177,
131
- 0.03751674000000094,
132
- 0.038046081000004506,
133
- 0.0378514770000038,
134
- 0.03802572299997564,
135
- 0.037510698000005505,
136
- 0.03811543099999426,
137
- 0.03790593900004069,
138
- 0.03746882999996615,
139
- 0.03797440799996821,
140
- 0.03935893000004853,
141
- 0.03864117599999872,
142
- 0.03823546600000327,
143
- 0.037051768000026186,
144
- 0.03772486900004424,
145
- 0.03833110499999748
 
 
 
 
146
  ]
147
  },
148
  "throughput": {
149
  "unit": "samples/s",
150
- "value": 26.850142961793406
151
  },
152
  "energy": {
153
  "unit": "kWh",
154
- "cpu": 1.2381615324152842e-06,
155
- "ram": 5.1744711971067524e-08,
156
  "gpu": 0.0,
157
- "total": 1.2899062443863518e-06
158
  },
159
  "efficiency": {
160
  "unit": "samples/kWh",
161
- "value": 775250.1426766338
162
  }
163
  }
164
  }
 
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.0",
83
+ "optimum_benchmark_commit": "70875ef900aa5b32d2b7eeaedc4a76bc99f650db",
84
  "transformers_version": "4.40.2",
85
  "transformers_commit": null,
86
  "accelerate_version": "0.30.1",
 
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 938.139648,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
107
  },
108
  "latency": {
109
  "unit": "s",
110
+ "count": 31,
111
+ "total": 1.0141728780001245,
112
+ "mean": 0.032715254129036274,
113
+ "stdev": 0.0005976728100066315,
114
+ "p50": 0.03255317099996091,
115
+ "p90": 0.033625195000013264,
116
+ "p95": 0.03367818449999049,
117
+ "p99": 0.03439317479998181,
118
  "values": [
119
+ 0.032726695999997446,
120
+ 0.032308700999976736,
121
+ 0.032540656000037416,
122
+ 0.03221091699998624,
123
+ 0.031932845000028465,
124
+ 0.03255317099996091,
125
+ 0.03267785499997444,
126
+ 0.03217025100002502,
127
+ 0.03264928100003317,
128
+ 0.0323256430000356,
129
+ 0.032312969000031444,
130
+ 0.032384815000000344,
131
+ 0.03197832000000744,
132
+ 0.03207892900002207,
133
+ 0.033625195000013264,
134
+ 0.03365021199999774,
135
+ 0.0333262440000226,
136
+ 0.03331711600003473,
137
+ 0.03315819699997746,
138
+ 0.03266425000003892,
139
+ 0.03254482500000222,
140
+ 0.032110228000021834,
141
+ 0.03242504999997209,
142
+ 0.032388390999983585,
143
+ 0.032984021000004304,
144
+ 0.03468761099998119,
145
+ 0.03294453500001282,
146
+ 0.03271607699997503,
147
+ 0.03370615699998325,
148
+ 0.03241823700000168,
149
+ 0.032655482999985
150
  ]
151
  },
152
  "throughput": {
153
  "unit": "samples/s",
154
+ "value": 30.566780745635555
155
  },
156
  "energy": {
157
  "unit": "kWh",
158
+ "cpu": 1.2507944057385129e-06,
159
+ "ram": 5.2272829605722514e-08,
160
  "gpu": 0.0,
161
+ "total": 1.3030672353442354e-06
162
  },
163
  "efficiency": {
164
  "unit": "samples/kWh",
165
+ "value": 767420.1091671428
166
  }
167
  }
168
  }