IlyasMoutawwakil HF staff commited on
Commit
0bbb236
1 Parent(s): caf9e4b

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cpu_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -73,23 +73,23 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 4,
76
- "cpu_ram_mb": 16757.354496,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
- "optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
84
- "transformers_version": "4.41.1",
85
  "transformers_commit": null,
86
- "accelerate_version": "0.30.1",
87
  "accelerate_commit": null,
88
- "diffusers_version": "0.27.2",
89
  "diffusers_commit": null,
90
  "optimum_version": null,
91
  "optimum_commit": null,
92
- "timm_version": "1.0.3",
93
  "timm_commit": null,
94
  "peft_version": null,
95
  "peft_commit": null
@@ -99,7 +99,7 @@
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
- "max_ram": 936.054784,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
@@ -107,60 +107,58 @@
107
  },
108
  "latency": {
109
  "unit": "s",
110
- "count": 29,
111
- "total": 1.0141411759996686,
112
- "mean": 0.03497038537929892,
113
- "stdev": 0.002453881649819887,
114
- "p50": 0.03616440100000773,
115
- "p90": 0.0372171237999737,
116
- "p95": 0.037343565599974225,
117
- "p99": 0.037539532639982554,
118
  "values": [
119
- 0.03298635800001648,
120
- 0.03694017800000893,
121
- 0.03681258199998183,
122
- 0.03532652699999517,
123
- 0.03712812699995993,
124
- 0.03675805099999252,
125
- 0.03616440100000773,
126
- 0.03593320199996697,
127
- 0.037354465999953845,
128
- 0.03761150299999372,
129
- 0.03718960099996593,
130
- 0.03608721699998796,
131
- 0.03669924199999741,
132
- 0.03557756300000392,
133
- 0.0373272150000048,
134
- 0.03677862899996853,
135
- 0.03686919699998725,
136
- 0.03660478699998748,
137
- 0.036206597999978385,
138
- 0.03680580899998631,
139
- 0.03490760900001533,
140
- 0.031344173999968916,
141
- 0.03185995000001185,
142
- 0.030997880999962035,
143
- 0.031525678999969386,
144
- 0.03126850400002468,
145
- 0.030910769000001892,
146
- 0.031019280999998955,
147
- 0.031146075999970435
148
  ]
149
  },
150
  "throughput": {
151
  "unit": "samples/s",
152
- "value": 28.595624244734815
153
  },
154
  "energy": {
155
  "unit": "kWh",
156
- "cpu": 1.2597496791552473e-06,
157
- "ram": 5.264666584881091e-08,
158
  "gpu": 0.0,
159
- "total": 1.312396345004058e-06
160
  },
161
  "efficiency": {
162
  "unit": "samples/kWh",
163
- "value": 761964.9382648257
164
  }
165
  }
166
  }
 
3
  "name": "cpu_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 4,
76
+ "cpu_ram_mb": 16757.338112,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-6.5.0-1022-azure-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.14",
82
  "optimum_benchmark_version": "0.2.1",
83
+ "optimum_benchmark_commit": "d920fe9626db1e7915f6d3574b5b54b0159cd100",
84
+ "transformers_version": "4.42.3",
85
  "transformers_commit": null,
86
+ "accelerate_version": "0.31.0",
87
  "accelerate_commit": null,
88
+ "diffusers_version": "0.29.2",
89
  "diffusers_commit": null,
90
  "optimum_version": null,
91
  "optimum_commit": null,
92
+ "timm_version": "1.0.7",
93
  "timm_commit": null,
94
  "peft_version": null,
95
  "peft_commit": null
 
99
  "forward": {
100
  "memory": {
101
  "unit": "MB",
102
+ "max_ram": 938.00448,
103
  "max_global_vram": null,
104
  "max_process_vram": null,
105
  "max_reserved": null,
 
107
  },
108
  "latency": {
109
  "unit": "s",
110
+ "count": 27,
111
+ "total": 1.0046809730000632,
112
+ "mean": 0.037210406407409746,
113
+ "stdev": 0.002002534550941035,
114
+ "p50": 0.038040260999991915,
115
+ "p90": 0.038528022599996346,
116
+ "p95": 0.038848303800011766,
117
+ "p99": 0.03899213389997613,
118
  "values": [
119
+ 0.03901360599996906,
120
+ 0.03830345199997964,
121
+ 0.03806761200002029,
122
+ 0.03713569399997141,
123
+ 0.038931020999996235,
124
+ 0.03806223200001568,
125
+ 0.03805387600004906,
126
+ 0.03833944899997732,
127
+ 0.03631287000001748,
128
+ 0.038417584999990595,
129
+ 0.03797265500003277,
130
+ 0.037760379000019384,
131
+ 0.03783275300003197,
132
+ 0.03726008600000341,
133
+ 0.038443172999961917,
134
+ 0.03701266399997394,
135
+ 0.038040260999991915,
136
+ 0.03839176700000735,
137
+ 0.037476269000023876,
138
+ 0.03781574199996385,
139
+ 0.0384110230000374,
140
+ 0.038655297000047995,
141
+ 0.038344537999989825,
142
+ 0.03384209299997565,
143
+ 0.03251436499999727,
144
+ 0.03232303000004322,
145
+ 0.03194748099997469
 
 
146
  ]
147
  },
148
  "throughput": {
149
  "unit": "samples/s",
150
+ "value": 26.8742025833093
151
  },
152
  "energy": {
153
  "unit": "kWh",
154
+ "cpu": 1.2830964553313447e-06,
155
+ "ram": 5.3622381735003835e-08,
156
  "gpu": 0.0,
157
+ "total": 1.3367188370663486e-06
158
  },
159
  "efficiency": {
160
  "unit": "samples/kWh",
161
+ "value": 748100.4772811207
162
  }
163
  }
164
  }