IlyasMoutawwakil HF staff commited on
Commit
18af712
1 Parent(s): f4d16e3

Upload cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cpu_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 939.343872,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
@@ -10,63 +10,62 @@
10
  },
11
  "latency": {
12
  "unit": "s",
13
- "count": 32,
14
- "total": 1.030907060000061,
15
- "mean": 0.032215845625001904,
16
- "stdev": 0.0007459957801647178,
17
- "p50": 0.032048246499982724,
18
- "p90": 0.032668799700019235,
19
- "p95": 0.03299822035002649,
20
- "p99": 0.03506179639999971,
21
  "values": [
22
- 0.03254714399997738,
23
- 0.03189269099999592,
24
- 0.03211841300003471,
25
- 0.032682317000023886,
26
- 0.03163615099998651,
27
- 0.032412109999995664,
28
- 0.03268702600001916,
29
- 0.03194115100001227,
30
- 0.032197952000046826,
31
- 0.03240295499995227,
32
- 0.031988269999999375,
33
- 0.03171021999997947,
34
- 0.03166960400000107,
35
- 0.031991805999950884,
36
- 0.03215371999999661,
37
- 0.03201253400004589,
38
- 0.03203597899999977,
39
- 0.0316945399999895,
40
- 0.03166980400004604,
41
- 0.03210710199999767,
42
- 0.03202093100003367,
43
- 0.03171705199997632,
44
- 0.03206051399996568,
45
- 0.032262532000004285,
46
- 0.032179246999987754,
47
- 0.03189115899999706,
48
- 0.03172356499999296,
49
- 0.03207491200004142,
50
- 0.03171822400003066,
51
- 0.03251083699996116,
52
- 0.03581802899998365,
53
- 0.03337856900003544
54
  ]
55
  },
56
  "throughput": {
57
  "unit": "samples/s",
58
- "value": 31.040625524475608
59
  },
60
  "energy": {
61
  "unit": "kWh",
62
- "cpu": 1.2249529361724854e-06,
63
- "ram": 5.119280590848577e-08,
64
  "gpu": 0.0,
65
- "total": 1.2761457420809711e-06
66
  },
67
  "efficiency": {
68
  "unit": "samples/kWh",
69
- "value": 783609.5572981588
70
  }
71
  }
72
  }
 
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 938.573824,
6
  "max_global_vram": null,
7
  "max_process_vram": null,
8
  "max_reserved": null,
 
10
  },
11
  "latency": {
12
  "unit": "s",
13
+ "count": 31,
14
+ "total": 1.0093720209999333,
15
+ "mean": 0.032560387774191395,
16
+ "stdev": 0.0026442566016224114,
17
+ "p50": 0.03157372399999758,
18
+ "p90": 0.03498879799997212,
19
+ "p95": 0.035783002499982786,
20
+ "p99": 0.04214882829998032,
21
  "values": [
22
+ 0.03284868999998025,
23
+ 0.031955966999987595,
24
+ 0.03125493999999662,
25
+ 0.03136103800000001,
26
+ 0.031235413999979755,
27
+ 0.03153669499999978,
28
+ 0.031188777000011214,
29
+ 0.03466599699999051,
30
+ 0.03163828399999602,
31
+ 0.03125595199998088,
32
+ 0.030857438999987608,
33
+ 0.03126983800001426,
34
+ 0.0316963729999884,
35
+ 0.031330550999996376,
36
+ 0.03157372399999758,
37
+ 0.03146095300002116,
38
+ 0.03175292800000307,
39
+ 0.03140262499999835,
40
+ 0.0321187100000202,
41
+ 0.03191768400000683,
42
+ 0.033956543999977384,
43
+ 0.04466513799997074,
44
+ 0.035288565999962884,
45
+ 0.03498879799997212,
46
+ 0.03627743900000269,
47
+ 0.03212666500002115,
48
+ 0.030946575000029952,
49
+ 0.030814800000030118,
50
+ 0.030479705000004742,
51
+ 0.03133616200000233,
52
+ 0.0341690500000027
 
53
  ]
54
  },
55
  "throughput": {
56
  "unit": "samples/s",
57
+ "value": 30.712164945180355
58
  },
59
  "energy": {
60
  "unit": "kWh",
61
+ "cpu": 1.1684858331493303e-06,
62
+ "ram": 4.883338134053407e-08,
63
  "gpu": 0.0,
64
+ "total": 1.2173192144898644e-06
65
  },
66
  "efficiency": {
67
  "unit": "samples/kWh",
68
+ "value": 821477.2165730292
69
  }
70
  }
71
  }