hf-transformers-bot commited on
Commit
dba3bd4
·
verified ·
1 Parent(s): 6e896e6

Upload folder using huggingface_hub

Browse files
Files changed (32) hide show
  1. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  2. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  3. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  4. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  5. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  6. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  7. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  8. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  9. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  10. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  11. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  12. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +2 -0
  13. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  14. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml +93 -0
  15. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml +141 -0
  16. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml +3 -0
  17. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json +473 -0
  18. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log +0 -0
  19. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json +104 -0
  20. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json +367 -0
  21. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json +13 -0
  22. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml +93 -0
  23. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml +141 -0
  24. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml +3 -0
  25. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json +473 -0
  26. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log +0 -0
  27. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json +104 -0
  28. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json +367 -0
  29. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json +13 -0
  30. 2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/multirun.yaml +234 -0
  31. 2024-09-16/summaries.json +41 -0
  32. 2024-09-16/summary.json +40 -0
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 8bd2b1e8c23234cd607ca8d63f53c1edfea27462
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
88
+ - hydra.run.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '0'
99
+ num: 0
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=False
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1333.57568,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 11.43847265625,
120
+ "mean": 11.43847265625,
121
+ "stdev": 0.0,
122
+ "p50": 11.43847265625,
123
+ "p90": 11.43847265625,
124
+ "p95": 11.43847265625,
125
+ "p99": 11.43847265625,
126
+ "values": [
127
+ 11.43847265625
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1797.885952,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.431872
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.04255776023864746,
147
+ "mean": 0.02127888011932373,
148
+ "stdev": 0.0005032320022583005,
149
+ "p50": 0.02127888011932373,
150
+ "p90": 0.02168146572113037,
151
+ "p95": 0.0217317889213562,
152
+ "p99": 0.021772047481536865,
153
+ "values": [
154
+ 0.02077564811706543,
155
+ 0.021782112121582032
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 328.964680506996
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1855.840256,
169
+ "max_global_vram": 6800.539648,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.801344
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.095236083984375,
178
+ "mean": 2.5476180419921874,
179
+ "stdev": 0.006811401367187386,
180
+ "p50": 2.5476180419921874,
181
+ "p90": 2.553067163085937,
182
+ "p95": 2.553748303222656,
183
+ "p99": 2.5542932153320312,
184
+ "values": [
185
+ 2.540806640625,
186
+ 2.554429443359375
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 49.85048696730397
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.094915065765382,
202
+ "mean": 0.020058720731359767,
203
+ "stdev": 0.0004947266597882792,
204
+ "p50": 0.020282367706298828,
205
+ "p90": 0.020450713729858398,
206
+ "p95": 0.020565453052520753,
207
+ "p99": 0.02098367437362671,
208
+ "values": [
209
+ 0.01925836753845215,
210
+ 0.01909350395202637,
211
+ 0.01986457633972168,
212
+ 0.0202926082611084,
213
+ 0.020384767532348632,
214
+ 0.020315135955810547,
215
+ 0.02031718444824219,
216
+ 0.02030080032348633,
217
+ 0.02026393508911133,
218
+ 0.020387840270996094,
219
+ 0.0204902400970459,
220
+ 0.020295679092407228,
221
+ 0.02025164794921875,
222
+ 0.02023423957824707,
223
+ 0.02030284881591797,
224
+ 0.020280319213867186,
225
+ 0.02028339195251465,
226
+ 0.020306943893432617,
227
+ 0.020152320861816408,
228
+ 0.020307968139648438,
229
+ 0.020262943267822266,
230
+ 0.02032431983947754,
231
+ 0.02104934310913086,
232
+ 0.020336639404296874,
233
+ 0.02025984001159668,
234
+ 0.020336639404296874,
235
+ 0.020583423614501953,
236
+ 0.020584447860717774,
237
+ 0.020288511276245116,
238
+ 0.02031001663208008,
239
+ 0.02031718444824219,
240
+ 0.020274175643920898,
241
+ 0.02026803207397461,
242
+ 0.020288511276245116,
243
+ 0.02031001663208008,
244
+ 0.020164608001708984,
245
+ 0.020281343460083007,
246
+ 0.02024448013305664,
247
+ 0.02027212715148926,
248
+ 0.02025881576538086,
249
+ 0.019319807052612305,
250
+ 0.01908531188964844,
251
+ 0.019094528198242186,
252
+ 0.01921433639526367,
253
+ 0.01927884864807129,
254
+ 0.019339263916015623,
255
+ 0.019306495666503908,
256
+ 0.019269632339477538,
257
+ 0.019320831298828126,
258
+ 0.019322944641113282,
259
+ 0.019286975860595704,
260
+ 0.019332096099853514,
261
+ 0.019508224487304687,
262
+ 0.01925119972229004,
263
+ 0.019302400588989257,
264
+ 0.019340288162231444,
265
+ 0.019371007919311522,
266
+ 0.019458047866821288,
267
+ 0.019438592910766602,
268
+ 0.01945088005065918,
269
+ 0.01944780731201172,
270
+ 0.01946419143676758,
271
+ 0.01940377616882324,
272
+ 0.01941094398498535,
273
+ 0.01941094398498535,
274
+ 0.01943552017211914,
275
+ 0.01940787124633789,
276
+ 0.01946316719055176,
277
+ 0.019530752182006835,
278
+ 0.019491840362548828,
279
+ 0.01942527961730957,
280
+ 0.019337215423583985,
281
+ 0.01947238349914551,
282
+ 0.01943654441833496,
283
+ 0.01944576072692871,
284
+ 0.019430400848388672,
285
+ 0.01944063949584961,
286
+ 0.01945395278930664,
287
+ 0.01965260887145996,
288
+ 0.019812351226806642,
289
+ 0.019749887466430666,
290
+ 0.01947340774536133,
291
+ 0.019417087554931642,
292
+ 0.01947340774536133,
293
+ 0.019796991348266603,
294
+ 0.020435968399047853,
295
+ 0.02040012741088867,
296
+ 0.020414464950561522,
297
+ 0.020358144760131838,
298
+ 0.02026700782775879,
299
+ 0.020386816024780274,
300
+ 0.020357120513916017,
301
+ 0.02040729522705078,
302
+ 0.02043187141418457,
303
+ 0.02040012741088867,
304
+ 0.020385791778564453,
305
+ 0.020452352523803712,
306
+ 0.020397056579589845,
307
+ 0.020354047775268554,
308
+ 0.020487167358398437,
309
+ 0.020405248641967775,
310
+ 0.020396032333374024,
311
+ 0.020380672454833985,
312
+ 0.020396064758300782,
313
+ 0.020424671173095703,
314
+ 0.020420608520507814,
315
+ 0.020314111709594726,
316
+ 0.020374528884887694,
317
+ 0.02042572784423828,
318
+ 0.020420608520507814,
319
+ 0.02038374328613281,
320
+ 0.020444160461425782,
321
+ 0.020328447341918944,
322
+ 0.020349952697753908,
323
+ 0.02041651153564453,
324
+ 0.020403200149536133,
325
+ 0.020396032333374024,
326
+ 0.020413440704345705,
327
+ 0.02039910316467285,
328
+ 0.020360191345214843,
329
+ 0.02043289566040039,
330
+ 0.020427776336669923,
331
+ 0.020368383407592772,
332
+ 0.020345855712890625,
333
+ 0.020446207046508787,
334
+ 0.020398080825805662,
335
+ 0.02046463966369629,
336
+ 0.02023526382446289,
337
+ 0.01905971145629883,
338
+ 0.019386367797851564,
339
+ 0.019742719650268553,
340
+ 0.019289087295532227,
341
+ 0.01927987289428711,
342
+ 0.01926348876953125,
343
+ 0.019331071853637697,
344
+ 0.019277824401855468,
345
+ 0.01922047996520996,
346
+ 0.019272703170776367,
347
+ 0.01924198341369629,
348
+ 0.0192675838470459,
349
+ 0.01922867202758789,
350
+ 0.019301376342773437,
351
+ 0.019306495666503908,
352
+ 0.01928704071044922,
353
+ 0.019319807052612305,
354
+ 0.01925836753845215,
355
+ 0.019268608093261717,
356
+ 0.01922150421142578,
357
+ 0.019358720779418945,
358
+ 0.019319807052612305,
359
+ 0.019309568405151366,
360
+ 0.019326976776123047,
361
+ 0.019339263916015623,
362
+ 0.019353599548339845,
363
+ 0.020555776596069338,
364
+ 0.021371904373168944,
365
+ 0.020529151916503906,
366
+ 0.020323328018188477,
367
+ 0.020239360809326173,
368
+ 0.020257791519165038,
369
+ 0.020265983581542968,
370
+ 0.02030080032348633,
371
+ 0.020312063217163084,
372
+ 0.02026188850402832,
373
+ 0.020265983581542968,
374
+ 0.02020249557495117,
375
+ 0.02030899238586426,
376
+ 0.020306943893432617,
377
+ 0.02026803207397461,
378
+ 0.02021990394592285,
379
+ 0.02024345588684082,
380
+ 0.020197376251220703,
381
+ 0.02023219108581543,
382
+ 0.020280319213867186,
383
+ 0.020256767272949217,
384
+ 0.020230144500732423,
385
+ 0.020286527633666993,
386
+ 0.020632511138916017,
387
+ 0.020388864517211915,
388
+ 0.020183040618896485,
389
+ 0.020229120254516602,
390
+ 0.02031001663208008,
391
+ 0.020730880737304686,
392
+ 0.020501504898071288,
393
+ 0.020395008087158203,
394
+ 0.02042367935180664,
395
+ 0.020354047775268554,
396
+ 0.020398080825805662,
397
+ 0.020384767532348632,
398
+ 0.020410400390625,
399
+ 0.020454368591308593,
400
+ 0.02043187141418457,
401
+ 0.020406272888183592,
402
+ 0.020068351745605468,
403
+ 0.019985408782958985,
404
+ 0.020265983581542968,
405
+ 0.02039910316467285,
406
+ 0.02045132827758789,
407
+ 0.020396032333374024,
408
+ 0.02040729522705078,
409
+ 0.020427776336669923,
410
+ 0.020435968399047853,
411
+ 0.020389888763427736,
412
+ 0.02046771240234375,
413
+ 0.020387903213500976,
414
+ 0.02040415954589844,
415
+ 0.020118528366088868,
416
+ 0.020132863998413086,
417
+ 0.02042163276672363,
418
+ 0.02040729522705078,
419
+ 0.02040012741088867,
420
+ 0.020322303771972656,
421
+ 0.020378623962402344,
422
+ 0.020133888244628906,
423
+ 0.020092927932739257,
424
+ 0.02044108772277832,
425
+ 0.020758527755737305,
426
+ 0.020397056579589845,
427
+ 0.02040012741088867,
428
+ 0.02037555122375488,
429
+ 0.020921344757080077,
430
+ 0.0216627197265625,
431
+ 0.020620288848876952,
432
+ 0.02044927978515625,
433
+ 0.0206561279296875,
434
+ 0.020488191604614257,
435
+ 0.020510719299316405,
436
+ 0.020477951049804686,
437
+ 0.020410367965698242,
438
+ 0.019340288162231444,
439
+ 0.019292160034179686,
440
+ 0.020303871154785155,
441
+ 0.020214784622192384,
442
+ 0.020925439834594727,
443
+ 0.0196945915222168,
444
+ 0.019513343811035155,
445
+ 0.019467264175415038,
446
+ 0.019530752182006835,
447
+ 0.02020147132873535,
448
+ 0.02007756805419922,
449
+ 0.020306943893432617,
450
+ 0.020388864517211915,
451
+ 0.020370431900024414,
452
+ 0.02008064079284668,
453
+ 0.020060159683227538,
454
+ 0.02042982482910156,
455
+ 0.019962879180908204,
456
+ 0.02003865623474121,
457
+ 0.020057088851928712,
458
+ 0.02028339195251465,
459
+ 0.020380672454833985,
460
+ 0.020642816543579103,
461
+ 0.020414464950561522,
462
+ 0.020404224395751954
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 49.85362792536424
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1333.57568,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 11.43847265625,
15
+ "mean": 11.43847265625,
16
+ "stdev": 0.0,
17
+ "p50": 11.43847265625,
18
+ "p90": 11.43847265625,
19
+ "p95": 11.43847265625,
20
+ "p99": 11.43847265625,
21
+ "values": [
22
+ 11.43847265625
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1797.885952,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.431872
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.04255776023864746,
42
+ "mean": 0.02127888011932373,
43
+ "stdev": 0.0005032320022583005,
44
+ "p50": 0.02127888011932373,
45
+ "p90": 0.02168146572113037,
46
+ "p95": 0.0217317889213562,
47
+ "p99": 0.021772047481536865,
48
+ "values": [
49
+ 0.02077564811706543,
50
+ 0.021782112121582032
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 328.964680506996
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1855.840256,
64
+ "max_global_vram": 6800.539648,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.801344
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.095236083984375,
73
+ "mean": 2.5476180419921874,
74
+ "stdev": 0.006811401367187386,
75
+ "p50": 2.5476180419921874,
76
+ "p90": 2.553067163085937,
77
+ "p95": 2.553748303222656,
78
+ "p99": 2.5542932153320312,
79
+ "values": [
80
+ 2.540806640625,
81
+ 2.554429443359375
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 49.85048696730397
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.094915065765382,
97
+ "mean": 0.020058720731359767,
98
+ "stdev": 0.0004947266597882792,
99
+ "p50": 0.020282367706298828,
100
+ "p90": 0.020450713729858398,
101
+ "p95": 0.020565453052520753,
102
+ "p99": 0.02098367437362671,
103
+ "values": [
104
+ 0.01925836753845215,
105
+ 0.01909350395202637,
106
+ 0.01986457633972168,
107
+ 0.0202926082611084,
108
+ 0.020384767532348632,
109
+ 0.020315135955810547,
110
+ 0.02031718444824219,
111
+ 0.02030080032348633,
112
+ 0.02026393508911133,
113
+ 0.020387840270996094,
114
+ 0.0204902400970459,
115
+ 0.020295679092407228,
116
+ 0.02025164794921875,
117
+ 0.02023423957824707,
118
+ 0.02030284881591797,
119
+ 0.020280319213867186,
120
+ 0.02028339195251465,
121
+ 0.020306943893432617,
122
+ 0.020152320861816408,
123
+ 0.020307968139648438,
124
+ 0.020262943267822266,
125
+ 0.02032431983947754,
126
+ 0.02104934310913086,
127
+ 0.020336639404296874,
128
+ 0.02025984001159668,
129
+ 0.020336639404296874,
130
+ 0.020583423614501953,
131
+ 0.020584447860717774,
132
+ 0.020288511276245116,
133
+ 0.02031001663208008,
134
+ 0.02031718444824219,
135
+ 0.020274175643920898,
136
+ 0.02026803207397461,
137
+ 0.020288511276245116,
138
+ 0.02031001663208008,
139
+ 0.020164608001708984,
140
+ 0.020281343460083007,
141
+ 0.02024448013305664,
142
+ 0.02027212715148926,
143
+ 0.02025881576538086,
144
+ 0.019319807052612305,
145
+ 0.01908531188964844,
146
+ 0.019094528198242186,
147
+ 0.01921433639526367,
148
+ 0.01927884864807129,
149
+ 0.019339263916015623,
150
+ 0.019306495666503908,
151
+ 0.019269632339477538,
152
+ 0.019320831298828126,
153
+ 0.019322944641113282,
154
+ 0.019286975860595704,
155
+ 0.019332096099853514,
156
+ 0.019508224487304687,
157
+ 0.01925119972229004,
158
+ 0.019302400588989257,
159
+ 0.019340288162231444,
160
+ 0.019371007919311522,
161
+ 0.019458047866821288,
162
+ 0.019438592910766602,
163
+ 0.01945088005065918,
164
+ 0.01944780731201172,
165
+ 0.01946419143676758,
166
+ 0.01940377616882324,
167
+ 0.01941094398498535,
168
+ 0.01941094398498535,
169
+ 0.01943552017211914,
170
+ 0.01940787124633789,
171
+ 0.01946316719055176,
172
+ 0.019530752182006835,
173
+ 0.019491840362548828,
174
+ 0.01942527961730957,
175
+ 0.019337215423583985,
176
+ 0.01947238349914551,
177
+ 0.01943654441833496,
178
+ 0.01944576072692871,
179
+ 0.019430400848388672,
180
+ 0.01944063949584961,
181
+ 0.01945395278930664,
182
+ 0.01965260887145996,
183
+ 0.019812351226806642,
184
+ 0.019749887466430666,
185
+ 0.01947340774536133,
186
+ 0.019417087554931642,
187
+ 0.01947340774536133,
188
+ 0.019796991348266603,
189
+ 0.020435968399047853,
190
+ 0.02040012741088867,
191
+ 0.020414464950561522,
192
+ 0.020358144760131838,
193
+ 0.02026700782775879,
194
+ 0.020386816024780274,
195
+ 0.020357120513916017,
196
+ 0.02040729522705078,
197
+ 0.02043187141418457,
198
+ 0.02040012741088867,
199
+ 0.020385791778564453,
200
+ 0.020452352523803712,
201
+ 0.020397056579589845,
202
+ 0.020354047775268554,
203
+ 0.020487167358398437,
204
+ 0.020405248641967775,
205
+ 0.020396032333374024,
206
+ 0.020380672454833985,
207
+ 0.020396064758300782,
208
+ 0.020424671173095703,
209
+ 0.020420608520507814,
210
+ 0.020314111709594726,
211
+ 0.020374528884887694,
212
+ 0.02042572784423828,
213
+ 0.020420608520507814,
214
+ 0.02038374328613281,
215
+ 0.020444160461425782,
216
+ 0.020328447341918944,
217
+ 0.020349952697753908,
218
+ 0.02041651153564453,
219
+ 0.020403200149536133,
220
+ 0.020396032333374024,
221
+ 0.020413440704345705,
222
+ 0.02039910316467285,
223
+ 0.020360191345214843,
224
+ 0.02043289566040039,
225
+ 0.020427776336669923,
226
+ 0.020368383407592772,
227
+ 0.020345855712890625,
228
+ 0.020446207046508787,
229
+ 0.020398080825805662,
230
+ 0.02046463966369629,
231
+ 0.02023526382446289,
232
+ 0.01905971145629883,
233
+ 0.019386367797851564,
234
+ 0.019742719650268553,
235
+ 0.019289087295532227,
236
+ 0.01927987289428711,
237
+ 0.01926348876953125,
238
+ 0.019331071853637697,
239
+ 0.019277824401855468,
240
+ 0.01922047996520996,
241
+ 0.019272703170776367,
242
+ 0.01924198341369629,
243
+ 0.0192675838470459,
244
+ 0.01922867202758789,
245
+ 0.019301376342773437,
246
+ 0.019306495666503908,
247
+ 0.01928704071044922,
248
+ 0.019319807052612305,
249
+ 0.01925836753845215,
250
+ 0.019268608093261717,
251
+ 0.01922150421142578,
252
+ 0.019358720779418945,
253
+ 0.019319807052612305,
254
+ 0.019309568405151366,
255
+ 0.019326976776123047,
256
+ 0.019339263916015623,
257
+ 0.019353599548339845,
258
+ 0.020555776596069338,
259
+ 0.021371904373168944,
260
+ 0.020529151916503906,
261
+ 0.020323328018188477,
262
+ 0.020239360809326173,
263
+ 0.020257791519165038,
264
+ 0.020265983581542968,
265
+ 0.02030080032348633,
266
+ 0.020312063217163084,
267
+ 0.02026188850402832,
268
+ 0.020265983581542968,
269
+ 0.02020249557495117,
270
+ 0.02030899238586426,
271
+ 0.020306943893432617,
272
+ 0.02026803207397461,
273
+ 0.02021990394592285,
274
+ 0.02024345588684082,
275
+ 0.020197376251220703,
276
+ 0.02023219108581543,
277
+ 0.020280319213867186,
278
+ 0.020256767272949217,
279
+ 0.020230144500732423,
280
+ 0.020286527633666993,
281
+ 0.020632511138916017,
282
+ 0.020388864517211915,
283
+ 0.020183040618896485,
284
+ 0.020229120254516602,
285
+ 0.02031001663208008,
286
+ 0.020730880737304686,
287
+ 0.020501504898071288,
288
+ 0.020395008087158203,
289
+ 0.02042367935180664,
290
+ 0.020354047775268554,
291
+ 0.020398080825805662,
292
+ 0.020384767532348632,
293
+ 0.020410400390625,
294
+ 0.020454368591308593,
295
+ 0.02043187141418457,
296
+ 0.020406272888183592,
297
+ 0.020068351745605468,
298
+ 0.019985408782958985,
299
+ 0.020265983581542968,
300
+ 0.02039910316467285,
301
+ 0.02045132827758789,
302
+ 0.020396032333374024,
303
+ 0.02040729522705078,
304
+ 0.020427776336669923,
305
+ 0.020435968399047853,
306
+ 0.020389888763427736,
307
+ 0.02046771240234375,
308
+ 0.020387903213500976,
309
+ 0.02040415954589844,
310
+ 0.020118528366088868,
311
+ 0.020132863998413086,
312
+ 0.02042163276672363,
313
+ 0.02040729522705078,
314
+ 0.02040012741088867,
315
+ 0.020322303771972656,
316
+ 0.020378623962402344,
317
+ 0.020133888244628906,
318
+ 0.020092927932739257,
319
+ 0.02044108772277832,
320
+ 0.020758527755737305,
321
+ 0.020397056579589845,
322
+ 0.02040012741088867,
323
+ 0.02037555122375488,
324
+ 0.020921344757080077,
325
+ 0.0216627197265625,
326
+ 0.020620288848876952,
327
+ 0.02044927978515625,
328
+ 0.0206561279296875,
329
+ 0.020488191604614257,
330
+ 0.020510719299316405,
331
+ 0.020477951049804686,
332
+ 0.020410367965698242,
333
+ 0.019340288162231444,
334
+ 0.019292160034179686,
335
+ 0.020303871154785155,
336
+ 0.020214784622192384,
337
+ 0.020925439834594727,
338
+ 0.0196945915222168,
339
+ 0.019513343811035155,
340
+ 0.019467264175415038,
341
+ 0.019530752182006835,
342
+ 0.02020147132873535,
343
+ 0.02007756805419922,
344
+ 0.020306943893432617,
345
+ 0.020388864517211915,
346
+ 0.020370431900024414,
347
+ 0.02008064079284668,
348
+ 0.020060159683227538,
349
+ 0.02042982482910156,
350
+ 0.019962879180908204,
351
+ 0.02003865623474121,
352
+ 0.020057088851928712,
353
+ 0.02028339195251465,
354
+ 0.020380672454833985,
355
+ 0.020642816543579103,
356
+ 0.020414464950561522,
357
+ 0.020404224395751954
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 49.85362792536424
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
4
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.02127888011932373,
7
+ "prefill.throughput.value": 328.964680506996,
8
+ "decode.latency.mean": 2.5476180419921874,
9
+ "decode.throughput.value": 49.85048696730397,
10
+ "per_token.latency.mean": 0.020058720731359767,
11
+ "per_token.throughput.value": 49.85362792536424
12
+ }
13
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: null
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 8bd2b1e8c23234cd607ca8d63f53c1edfea27462
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
88
+ - hydra.run.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '1'
99
+ num: 1
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=null
3
+ - backend.torch_compile=True
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [ISOLATED-PROCESS][2024-09-16 02:49:09,550][process][ERROR] - + Sending traceback to main process
2
+ [MAIN-PROCESS][2024-09-16 02:49:11,982][process][ERROR] - + Received traceback from isolated process
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=null,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": null,
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: false
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 8bd2b1e8c23234cd607ca8d63f53c1edfea27462
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
88
+ - hydra.run.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=False
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
98
+ id: '2'
99
+ num: 2
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=False
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1334.009856,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 11.40590625,
120
+ "mean": 11.40590625,
121
+ "stdev": 0.0,
122
+ "p50": 11.40590625,
123
+ "p90": 11.40590625,
124
+ "p95": 11.40590625,
125
+ "p99": 11.40590625,
126
+ "values": [
127
+ 11.40590625
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1848.99584,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6142.558208,
141
+ "max_allocated": 5028.450816
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.045700126647949216,
147
+ "mean": 0.022850063323974608,
148
+ "stdev": 5.508804321289161e-05,
149
+ "p50": 0.022850063323974608,
150
+ "p90": 0.022894133758544922,
151
+ "p95": 0.022899642562866213,
152
+ "p99": 0.022904049606323244,
153
+ "values": [
154
+ 0.022794975280761718,
155
+ 0.0229051513671875
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 306.3448840710871
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 1875.365888,
169
+ "max_global_vram": 6800.539648,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 6146.752512,
172
+ "max_allocated": 5031.820288
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 5.383725097656249,
178
+ "mean": 2.6918625488281247,
179
+ "stdev": 0.00799218749999997,
180
+ "p50": 2.6918625488281247,
181
+ "p90": 2.698256298828125,
182
+ "p95": 2.699055517578125,
183
+ "p99": 2.699694892578125,
184
+ "values": [
185
+ 2.699854736328125,
186
+ 2.683870361328125
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 47.17922913831101
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 5.383459831237792,
202
+ "mean": 0.02119472374503068,
203
+ "stdev": 0.00028502548569978985,
204
+ "p50": 0.02109183979034424,
205
+ "p90": 0.021576192474365236,
206
+ "p95": 0.021695386123657227,
207
+ "p99": 0.02194905113220215,
208
+ "values": [
209
+ 0.02110771179199219,
210
+ 0.021518335342407227,
211
+ 0.02109951972961426,
212
+ 0.02103603172302246,
213
+ 0.02110054397583008,
214
+ 0.02104934310913086,
215
+ 0.02105855941772461,
216
+ 0.021006336212158205,
217
+ 0.02105446434020996,
218
+ 0.02108006477355957,
219
+ 0.021061632156372072,
220
+ 0.021098495483398438,
221
+ 0.021188608169555666,
222
+ 0.020997119903564454,
223
+ 0.02105241584777832,
224
+ 0.02101759910583496,
225
+ 0.02109951972961426,
226
+ 0.021169151306152344,
227
+ 0.021353471755981446,
228
+ 0.02106060791015625,
229
+ 0.02110771179199219,
230
+ 0.02103398323059082,
231
+ 0.02109132766723633,
232
+ 0.02102681541442871,
233
+ 0.020966400146484376,
234
+ 0.021206016540527343,
235
+ 0.0211015682220459,
236
+ 0.021078016281127928,
237
+ 0.0210513916015625,
238
+ 0.021200895309448242,
239
+ 0.02168524742126465,
240
+ 0.02172211265563965,
241
+ 0.02187775993347168,
242
+ 0.0213309440612793,
243
+ 0.021367807388305664,
244
+ 0.0210831356048584,
245
+ 0.02103603172302246,
246
+ 0.021922815322875978,
247
+ 0.021642240524291992,
248
+ 0.021142528533935546,
249
+ 0.02091929626464844,
250
+ 0.021207040786743164,
251
+ 0.021152767181396484,
252
+ 0.02110361671447754,
253
+ 0.02109132766723633,
254
+ 0.02169036865234375,
255
+ 0.021153791427612305,
256
+ 0.021704704284667968,
257
+ 0.021638143539428712,
258
+ 0.021974016189575195,
259
+ 0.021114879608154297,
260
+ 0.0214968318939209,
261
+ 0.021097471237182617,
262
+ 0.02107084846496582,
263
+ 0.02110054397583008,
264
+ 0.02104012870788574,
265
+ 0.021097471237182617,
266
+ 0.02103910446166992,
267
+ 0.02106470489501953,
268
+ 0.02103500747680664,
269
+ 0.02126131248474121,
270
+ 0.021571584701538086,
271
+ 0.02143948745727539,
272
+ 0.021123071670532227,
273
+ 0.021129215240478515,
274
+ 0.021390335083007812,
275
+ 0.02106879997253418,
276
+ 0.021105663299560547,
277
+ 0.021283840179443358,
278
+ 0.02165862464904785,
279
+ 0.021740543365478517,
280
+ 0.021577728271484374,
281
+ 0.021741567611694337,
282
+ 0.020960256576538085,
283
+ 0.02106060791015625,
284
+ 0.02105036735534668,
285
+ 0.02108723258972168,
286
+ 0.02102272033691406,
287
+ 0.02109542465209961,
288
+ 0.021586944580078125,
289
+ 0.021514240264892577,
290
+ 0.021007360458374022,
291
+ 0.02104934310913086,
292
+ 0.021082111358642578,
293
+ 0.021342208862304687,
294
+ 0.02104832077026367,
295
+ 0.021369855880737306,
296
+ 0.021501951217651367,
297
+ 0.02153267288208008,
298
+ 0.02103193664550781,
299
+ 0.021053440093994142,
300
+ 0.02213580894470215,
301
+ 0.021793792724609375,
302
+ 0.021157888412475585,
303
+ 0.021128192901611328,
304
+ 0.021118976593017577,
305
+ 0.02147737693786621,
306
+ 0.021420032501220702,
307
+ 0.02150297546386719,
308
+ 0.02150912094116211,
309
+ 0.021409791946411134,
310
+ 0.02106675148010254,
311
+ 0.021343231201171875,
312
+ 0.02105855941772461,
313
+ 0.021097471237182617,
314
+ 0.020995199203491213,
315
+ 0.020960128784179688,
316
+ 0.021177343368530274,
317
+ 0.021380096435546874,
318
+ 0.021489664077758788,
319
+ 0.0210831356048584,
320
+ 0.02142617607116699,
321
+ 0.021515264511108398,
322
+ 0.021485567092895508,
323
+ 0.02147020721435547,
324
+ 0.021596160888671875,
325
+ 0.021596160888671875,
326
+ 0.021515264511108398,
327
+ 0.021130239486694336,
328
+ 0.021597183227539063,
329
+ 0.021045248031616212,
330
+ 0.02100223922729492,
331
+ 0.021402624130249022,
332
+ 0.02148249626159668,
333
+ 0.020998144149780275,
334
+ 0.02099404716491699,
335
+ 0.02099404716491699,
336
+ 0.02101862335205078,
337
+ 0.021572608947753907,
338
+ 0.021321727752685548,
339
+ 0.021083200454711914,
340
+ 0.021033920288085938,
341
+ 0.02104934310913086,
342
+ 0.02103091239929199,
343
+ 0.02147532844543457,
344
+ 0.021032960891723632,
345
+ 0.021012479782104493,
346
+ 0.02100223922729492,
347
+ 0.020973567962646485,
348
+ 0.02106368064880371,
349
+ 0.021014528274536134,
350
+ 0.021790719985961913,
351
+ 0.021028863906860353,
352
+ 0.021037055969238282,
353
+ 0.02103603172302246,
354
+ 0.020998144149780275,
355
+ 0.02122137641906738,
356
+ 0.021102592468261717,
357
+ 0.02168627166748047,
358
+ 0.021511167526245118,
359
+ 0.021346303939819337,
360
+ 0.021011455535888672,
361
+ 0.020974592208862306,
362
+ 0.020998207092285157,
363
+ 0.02121516799926758,
364
+ 0.021926912307739257,
365
+ 0.02369638442993164,
366
+ 0.021772287368774415,
367
+ 0.021203968048095705,
368
+ 0.02105958366394043,
369
+ 0.02105753517150879,
370
+ 0.021004352569580078,
371
+ 0.021076927185058592,
372
+ 0.021585920333862304,
373
+ 0.02166783905029297,
374
+ 0.021213184356689452,
375
+ 0.021106687545776368,
376
+ 0.02109542465209961,
377
+ 0.0211015682220459,
378
+ 0.021103679656982424,
379
+ 0.021008319854736328,
380
+ 0.020995071411132812,
381
+ 0.02102783966064453,
382
+ 0.02104422378540039,
383
+ 0.021073919296264648,
384
+ 0.0210565128326416,
385
+ 0.02104115104675293,
386
+ 0.02105241584777832,
387
+ 0.021014528274536134,
388
+ 0.020993024826049804,
389
+ 0.02106982421875,
390
+ 0.021014528274536134,
391
+ 0.021008384704589843,
392
+ 0.02107699203491211,
393
+ 0.02104934310913086,
394
+ 0.02108723258972168,
395
+ 0.021123071670532227,
396
+ 0.02086604881286621,
397
+ 0.021398527145385742,
398
+ 0.021097471237182617,
399
+ 0.021001216888427734,
400
+ 0.021160959243774414,
401
+ 0.021113855361938477,
402
+ 0.021090303421020508,
403
+ 0.021131263732910157,
404
+ 0.020875263214111327,
405
+ 0.02084351921081543,
406
+ 0.02084864044189453,
407
+ 0.02102272033691406,
408
+ 0.02106368064880371,
409
+ 0.02105753517150879,
410
+ 0.02111692810058594,
411
+ 0.02106265640258789,
412
+ 0.0210565128326416,
413
+ 0.02105753517150879,
414
+ 0.02105855941772461,
415
+ 0.020943872451782225,
416
+ 0.021034015655517577,
417
+ 0.021056480407714843,
418
+ 0.02104832077026367,
419
+ 0.021127168655395507,
420
+ 0.02109951972961426,
421
+ 0.02107084846496582,
422
+ 0.02104012870788574,
423
+ 0.021073919296264648,
424
+ 0.02110054397583008,
425
+ 0.02108415985107422,
426
+ 0.0210882568359375,
427
+ 0.021098495483398438,
428
+ 0.021166080474853514,
429
+ 0.021168127059936523,
430
+ 0.021161983489990235,
431
+ 0.02106166458129883,
432
+ 0.02093667221069336,
433
+ 0.021105663299560547,
434
+ 0.021086208343505858,
435
+ 0.02109644889831543,
436
+ 0.02105958366394043,
437
+ 0.02108518409729004,
438
+ 0.021127168655395507,
439
+ 0.02109132766723633,
440
+ 0.021166080474853514,
441
+ 0.021061632156372072,
442
+ 0.02110873603820801,
443
+ 0.021304319381713867,
444
+ 0.021082111358642578,
445
+ 0.0210565128326416,
446
+ 0.02148863983154297,
447
+ 0.021113855361938477,
448
+ 0.020939775466918945,
449
+ 0.02110771179199219,
450
+ 0.02109235191345215,
451
+ 0.021126144409179686,
452
+ 0.02107494354248047,
453
+ 0.021073919296264648,
454
+ 0.02106265640258789,
455
+ 0.021123071670532227,
456
+ 0.02083328056335449,
457
+ 0.02085683250427246,
458
+ 0.020840448379516603,
459
+ 0.02106060791015625,
460
+ 0.02109951972961426,
461
+ 0.02104319953918457,
462
+ 0.02103500747680664
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 47.18155386358646
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark.log ADDED
File without changes
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": false,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1334.009856,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 11.40590625,
15
+ "mean": 11.40590625,
16
+ "stdev": 0.0,
17
+ "p50": 11.40590625,
18
+ "p90": 11.40590625,
19
+ "p95": 11.40590625,
20
+ "p99": 11.40590625,
21
+ "values": [
22
+ 11.40590625
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1848.99584,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6142.558208,
36
+ "max_allocated": 5028.450816
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.045700126647949216,
42
+ "mean": 0.022850063323974608,
43
+ "stdev": 5.508804321289161e-05,
44
+ "p50": 0.022850063323974608,
45
+ "p90": 0.022894133758544922,
46
+ "p95": 0.022899642562866213,
47
+ "p99": 0.022904049606323244,
48
+ "values": [
49
+ 0.022794975280761718,
50
+ 0.0229051513671875
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 306.3448840710871
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 1875.365888,
64
+ "max_global_vram": 6800.539648,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 6146.752512,
67
+ "max_allocated": 5031.820288
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 5.383725097656249,
73
+ "mean": 2.6918625488281247,
74
+ "stdev": 0.00799218749999997,
75
+ "p50": 2.6918625488281247,
76
+ "p90": 2.698256298828125,
77
+ "p95": 2.699055517578125,
78
+ "p99": 2.699694892578125,
79
+ "values": [
80
+ 2.699854736328125,
81
+ 2.683870361328125
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 47.17922913831101
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 5.383459831237792,
97
+ "mean": 0.02119472374503068,
98
+ "stdev": 0.00028502548569978985,
99
+ "p50": 0.02109183979034424,
100
+ "p90": 0.021576192474365236,
101
+ "p95": 0.021695386123657227,
102
+ "p99": 0.02194905113220215,
103
+ "values": [
104
+ 0.02110771179199219,
105
+ 0.021518335342407227,
106
+ 0.02109951972961426,
107
+ 0.02103603172302246,
108
+ 0.02110054397583008,
109
+ 0.02104934310913086,
110
+ 0.02105855941772461,
111
+ 0.021006336212158205,
112
+ 0.02105446434020996,
113
+ 0.02108006477355957,
114
+ 0.021061632156372072,
115
+ 0.021098495483398438,
116
+ 0.021188608169555666,
117
+ 0.020997119903564454,
118
+ 0.02105241584777832,
119
+ 0.02101759910583496,
120
+ 0.02109951972961426,
121
+ 0.021169151306152344,
122
+ 0.021353471755981446,
123
+ 0.02106060791015625,
124
+ 0.02110771179199219,
125
+ 0.02103398323059082,
126
+ 0.02109132766723633,
127
+ 0.02102681541442871,
128
+ 0.020966400146484376,
129
+ 0.021206016540527343,
130
+ 0.0211015682220459,
131
+ 0.021078016281127928,
132
+ 0.0210513916015625,
133
+ 0.021200895309448242,
134
+ 0.02168524742126465,
135
+ 0.02172211265563965,
136
+ 0.02187775993347168,
137
+ 0.0213309440612793,
138
+ 0.021367807388305664,
139
+ 0.0210831356048584,
140
+ 0.02103603172302246,
141
+ 0.021922815322875978,
142
+ 0.021642240524291992,
143
+ 0.021142528533935546,
144
+ 0.02091929626464844,
145
+ 0.021207040786743164,
146
+ 0.021152767181396484,
147
+ 0.02110361671447754,
148
+ 0.02109132766723633,
149
+ 0.02169036865234375,
150
+ 0.021153791427612305,
151
+ 0.021704704284667968,
152
+ 0.021638143539428712,
153
+ 0.021974016189575195,
154
+ 0.021114879608154297,
155
+ 0.0214968318939209,
156
+ 0.021097471237182617,
157
+ 0.02107084846496582,
158
+ 0.02110054397583008,
159
+ 0.02104012870788574,
160
+ 0.021097471237182617,
161
+ 0.02103910446166992,
162
+ 0.02106470489501953,
163
+ 0.02103500747680664,
164
+ 0.02126131248474121,
165
+ 0.021571584701538086,
166
+ 0.02143948745727539,
167
+ 0.021123071670532227,
168
+ 0.021129215240478515,
169
+ 0.021390335083007812,
170
+ 0.02106879997253418,
171
+ 0.021105663299560547,
172
+ 0.021283840179443358,
173
+ 0.02165862464904785,
174
+ 0.021740543365478517,
175
+ 0.021577728271484374,
176
+ 0.021741567611694337,
177
+ 0.020960256576538085,
178
+ 0.02106060791015625,
179
+ 0.02105036735534668,
180
+ 0.02108723258972168,
181
+ 0.02102272033691406,
182
+ 0.02109542465209961,
183
+ 0.021586944580078125,
184
+ 0.021514240264892577,
185
+ 0.021007360458374022,
186
+ 0.02104934310913086,
187
+ 0.021082111358642578,
188
+ 0.021342208862304687,
189
+ 0.02104832077026367,
190
+ 0.021369855880737306,
191
+ 0.021501951217651367,
192
+ 0.02153267288208008,
193
+ 0.02103193664550781,
194
+ 0.021053440093994142,
195
+ 0.02213580894470215,
196
+ 0.021793792724609375,
197
+ 0.021157888412475585,
198
+ 0.021128192901611328,
199
+ 0.021118976593017577,
200
+ 0.02147737693786621,
201
+ 0.021420032501220702,
202
+ 0.02150297546386719,
203
+ 0.02150912094116211,
204
+ 0.021409791946411134,
205
+ 0.02106675148010254,
206
+ 0.021343231201171875,
207
+ 0.02105855941772461,
208
+ 0.021097471237182617,
209
+ 0.020995199203491213,
210
+ 0.020960128784179688,
211
+ 0.021177343368530274,
212
+ 0.021380096435546874,
213
+ 0.021489664077758788,
214
+ 0.0210831356048584,
215
+ 0.02142617607116699,
216
+ 0.021515264511108398,
217
+ 0.021485567092895508,
218
+ 0.02147020721435547,
219
+ 0.021596160888671875,
220
+ 0.021596160888671875,
221
+ 0.021515264511108398,
222
+ 0.021130239486694336,
223
+ 0.021597183227539063,
224
+ 0.021045248031616212,
225
+ 0.02100223922729492,
226
+ 0.021402624130249022,
227
+ 0.02148249626159668,
228
+ 0.020998144149780275,
229
+ 0.02099404716491699,
230
+ 0.02099404716491699,
231
+ 0.02101862335205078,
232
+ 0.021572608947753907,
233
+ 0.021321727752685548,
234
+ 0.021083200454711914,
235
+ 0.021033920288085938,
236
+ 0.02104934310913086,
237
+ 0.02103091239929199,
238
+ 0.02147532844543457,
239
+ 0.021032960891723632,
240
+ 0.021012479782104493,
241
+ 0.02100223922729492,
242
+ 0.020973567962646485,
243
+ 0.02106368064880371,
244
+ 0.021014528274536134,
245
+ 0.021790719985961913,
246
+ 0.021028863906860353,
247
+ 0.021037055969238282,
248
+ 0.02103603172302246,
249
+ 0.020998144149780275,
250
+ 0.02122137641906738,
251
+ 0.021102592468261717,
252
+ 0.02168627166748047,
253
+ 0.021511167526245118,
254
+ 0.021346303939819337,
255
+ 0.021011455535888672,
256
+ 0.020974592208862306,
257
+ 0.020998207092285157,
258
+ 0.02121516799926758,
259
+ 0.021926912307739257,
260
+ 0.02369638442993164,
261
+ 0.021772287368774415,
262
+ 0.021203968048095705,
263
+ 0.02105958366394043,
264
+ 0.02105753517150879,
265
+ 0.021004352569580078,
266
+ 0.021076927185058592,
267
+ 0.021585920333862304,
268
+ 0.02166783905029297,
269
+ 0.021213184356689452,
270
+ 0.021106687545776368,
271
+ 0.02109542465209961,
272
+ 0.0211015682220459,
273
+ 0.021103679656982424,
274
+ 0.021008319854736328,
275
+ 0.020995071411132812,
276
+ 0.02102783966064453,
277
+ 0.02104422378540039,
278
+ 0.021073919296264648,
279
+ 0.0210565128326416,
280
+ 0.02104115104675293,
281
+ 0.02105241584777832,
282
+ 0.021014528274536134,
283
+ 0.020993024826049804,
284
+ 0.02106982421875,
285
+ 0.021014528274536134,
286
+ 0.021008384704589843,
287
+ 0.02107699203491211,
288
+ 0.02104934310913086,
289
+ 0.02108723258972168,
290
+ 0.021123071670532227,
291
+ 0.02086604881286621,
292
+ 0.021398527145385742,
293
+ 0.021097471237182617,
294
+ 0.021001216888427734,
295
+ 0.021160959243774414,
296
+ 0.021113855361938477,
297
+ 0.021090303421020508,
298
+ 0.021131263732910157,
299
+ 0.020875263214111327,
300
+ 0.02084351921081543,
301
+ 0.02084864044189453,
302
+ 0.02102272033691406,
303
+ 0.02106368064880371,
304
+ 0.02105753517150879,
305
+ 0.02111692810058594,
306
+ 0.02106265640258789,
307
+ 0.0210565128326416,
308
+ 0.02105753517150879,
309
+ 0.02105855941772461,
310
+ 0.020943872451782225,
311
+ 0.021034015655517577,
312
+ 0.021056480407714843,
313
+ 0.02104832077026367,
314
+ 0.021127168655395507,
315
+ 0.02109951972961426,
316
+ 0.02107084846496582,
317
+ 0.02104012870788574,
318
+ 0.021073919296264648,
319
+ 0.02110054397583008,
320
+ 0.02108415985107422,
321
+ 0.0210882568359375,
322
+ 0.021098495483398438,
323
+ 0.021166080474853514,
324
+ 0.021168127059936523,
325
+ 0.021161983489990235,
326
+ 0.02106166458129883,
327
+ 0.02093667221069336,
328
+ 0.021105663299560547,
329
+ 0.021086208343505858,
330
+ 0.02109644889831543,
331
+ 0.02105958366394043,
332
+ 0.02108518409729004,
333
+ 0.021127168655395507,
334
+ 0.02109132766723633,
335
+ 0.021166080474853514,
336
+ 0.021061632156372072,
337
+ 0.02110873603820801,
338
+ 0.021304319381713867,
339
+ 0.021082111358642578,
340
+ 0.0210565128326416,
341
+ 0.02148863983154297,
342
+ 0.021113855361938477,
343
+ 0.020939775466918945,
344
+ 0.02110771179199219,
345
+ 0.02109235191345215,
346
+ 0.021126144409179686,
347
+ 0.02107494354248047,
348
+ 0.021073919296264648,
349
+ 0.02106265640258789,
350
+ 0.021123071670532227,
351
+ 0.02083328056335449,
352
+ 0.02085683250427246,
353
+ 0.020840448379516603,
354
+ 0.02106060791015625,
355
+ 0.02109951972961426,
356
+ 0.02104319953918457,
357
+ 0.02103500747680664
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 47.18155386358646
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=False/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.022850063323974608,
7
+ "prefill.throughput.value": 306.3448840710871,
8
+ "decode.latency.mean": 2.6918625488281247,
9
+ "decode.throughput.value": 47.17922913831101,
10
+ "per_token.latency.mean": 0.02119472374503068,
11
+ "per_token.throughput.value": 47.18155386358646
12
+ }
13
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/config.yaml ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: pytorch_generate
2
+ backend:
3
+ name: pytorch
4
+ version: 2.4.0+cu121
5
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
6
+ task: null
7
+ library: null
8
+ model_type: null
9
+ model: google/gemma-2b
10
+ processor: null
11
+ device: cuda
12
+ device_ids: '0'
13
+ seed: 42
14
+ inter_op_num_threads: null
15
+ intra_op_num_threads: null
16
+ model_kwargs: {}
17
+ processor_kwargs: {}
18
+ no_weights: true
19
+ device_map: null
20
+ torch_dtype: float16
21
+ eval_mode: true
22
+ to_bettertransformer: false
23
+ low_cpu_mem_usage: null
24
+ attn_implementation: null
25
+ cache_implementation: static
26
+ autocast_enabled: false
27
+ autocast_dtype: null
28
+ torch_compile: true
29
+ torch_compile_target: forward
30
+ torch_compile_config:
31
+ backend: inductor
32
+ mode: reduce-overhead
33
+ fullgraph: true
34
+ quantization_scheme: null
35
+ quantization_config: {}
36
+ deepspeed_inference: false
37
+ deepspeed_inference_config: {}
38
+ peft_type: null
39
+ peft_config: {}
40
+ scenario:
41
+ name: inference
42
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
43
+ iterations: 2
44
+ duration: 0
45
+ warmup_runs: 10
46
+ input_shapes:
47
+ batch_size: 1
48
+ sequence_length: 7
49
+ new_tokens: null
50
+ memory: true
51
+ latency: true
52
+ energy: false
53
+ forward_kwargs: {}
54
+ generate_kwargs:
55
+ max_new_tokens: 128
56
+ min_new_tokens: 128
57
+ do_sample: false
58
+ call_kwargs: {}
59
+ launcher:
60
+ name: process
61
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
62
+ device_isolation: true
63
+ device_isolation_action: warn
64
+ numactl: false
65
+ numactl_kwargs: {}
66
+ start_method: spawn
67
+ environment:
68
+ cpu: ' AMD EPYC 7R32'
69
+ cpu_count: 16
70
+ cpu_ram_mb: 66697.261056
71
+ system: Linux
72
+ machine: x86_64
73
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
74
+ processor: x86_64
75
+ python_version: 3.8.10
76
+ gpu:
77
+ - NVIDIA A10G
78
+ gpu_count: 1
79
+ gpu_vram_mb: 24146608128
80
+ optimum_benchmark_version: 0.4.0
81
+ optimum_benchmark_commit: null
82
+ transformers_version: 4.45.0.dev0
83
+ transformers_commit: 8bd2b1e8c23234cd607ca8d63f53c1edfea27462
84
+ accelerate_version: 0.35.0.dev0
85
+ accelerate_commit: null
86
+ diffusers_version: null
87
+ diffusers_commit: null
88
+ optimum_version: 1.23.0.dev0
89
+ optimum_commit: null
90
+ timm_version: 0.9.16
91
+ timm_commit: null
92
+ peft_version: 0.12.1.dev0
93
+ peft_commit: null
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/hydra.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
88
+ - hydra.run.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=static
93
+ - backend.torch_compile=True
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
98
+ id: '3'
99
+ num: 3
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: /transformers/_benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/.hydra/overrides.yaml ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ - backend.model=google/gemma-2b
2
+ - backend.cache_implementation=static
3
+ - backend.torch_compile=True
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.json ADDED
@@ -0,0 +1,473 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "pytorch_generate",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "text-generation",
9
+ "library": "transformers",
10
+ "model_type": "gemma",
11
+ "model": "google/gemma-2b",
12
+ "processor": "google/gemma-2b",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": "float16",
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": "static",
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": true,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {
33
+ "backend": "inductor",
34
+ "mode": "reduce-overhead",
35
+ "fullgraph": true
36
+ },
37
+ "quantization_scheme": null,
38
+ "quantization_config": {},
39
+ "deepspeed_inference": false,
40
+ "deepspeed_inference_config": {},
41
+ "peft_type": null,
42
+ "peft_config": {}
43
+ },
44
+ "scenario": {
45
+ "name": "inference",
46
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
47
+ "iterations": 2,
48
+ "duration": 0,
49
+ "warmup_runs": 10,
50
+ "input_shapes": {
51
+ "batch_size": 1,
52
+ "num_choices": 2,
53
+ "sequence_length": 7
54
+ },
55
+ "new_tokens": null,
56
+ "memory": true,
57
+ "latency": true,
58
+ "energy": false,
59
+ "forward_kwargs": {},
60
+ "generate_kwargs": {
61
+ "max_new_tokens": 128,
62
+ "min_new_tokens": 128,
63
+ "do_sample": false
64
+ },
65
+ "call_kwargs": {}
66
+ },
67
+ "launcher": {
68
+ "name": "process",
69
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
70
+ "device_isolation": true,
71
+ "device_isolation_action": "warn",
72
+ "numactl": false,
73
+ "numactl_kwargs": {},
74
+ "start_method": "spawn"
75
+ },
76
+ "environment": {
77
+ "cpu": " AMD EPYC 7R32",
78
+ "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.261056,
80
+ "system": "Linux",
81
+ "machine": "x86_64",
82
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
83
+ "processor": "x86_64",
84
+ "python_version": "3.8.10",
85
+ "gpu": [
86
+ "NVIDIA A10G"
87
+ ],
88
+ "gpu_count": 1,
89
+ "gpu_vram_mb": 24146608128,
90
+ "optimum_benchmark_version": "0.4.0",
91
+ "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.45.0.dev0",
93
+ "transformers_commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
94
+ "accelerate_version": "0.35.0.dev0",
95
+ "accelerate_commit": null,
96
+ "diffusers_version": null,
97
+ "diffusers_commit": null,
98
+ "optimum_version": "1.23.0.dev0",
99
+ "optimum_commit": null,
100
+ "timm_version": "0.9.16",
101
+ "timm_commit": null,
102
+ "peft_version": "0.12.1.dev0",
103
+ "peft_commit": null
104
+ }
105
+ },
106
+ "report": {
107
+ "load": {
108
+ "memory": {
109
+ "unit": "MB",
110
+ "max_ram": 1353.68704,
111
+ "max_global_vram": 6775.373824,
112
+ "max_process_vram": 0.0,
113
+ "max_reserved": 6138.363904,
114
+ "max_allocated": 6060.931072
115
+ },
116
+ "latency": {
117
+ "unit": "s",
118
+ "count": 1,
119
+ "total": 12.495619140625,
120
+ "mean": 12.495619140625,
121
+ "stdev": 0.0,
122
+ "p50": 12.495619140625,
123
+ "p90": 12.495619140625,
124
+ "p95": 12.495619140625,
125
+ "p99": 12.495619140625,
126
+ "values": [
127
+ 12.495619140625
128
+ ]
129
+ },
130
+ "throughput": null,
131
+ "energy": null,
132
+ "efficiency": null
133
+ },
134
+ "prefill": {
135
+ "memory": {
136
+ "unit": "MB",
137
+ "max_ram": 1931.149312,
138
+ "max_global_vram": 6796.345344,
139
+ "max_process_vram": 0.0,
140
+ "max_reserved": 6138.363904,
141
+ "max_allocated": 5278.596608
142
+ },
143
+ "latency": {
144
+ "unit": "s",
145
+ "count": 2,
146
+ "total": 0.02882156753540039,
147
+ "mean": 0.014410783767700195,
148
+ "stdev": 0.00048147201538085993,
149
+ "p50": 0.014410783767700195,
150
+ "p90": 0.014795961380004883,
151
+ "p95": 0.014844108581542968,
152
+ "p99": 0.014882626342773438,
153
+ "values": [
154
+ 0.013929311752319335,
155
+ 0.014892255783081055
156
+ ]
157
+ },
158
+ "throughput": {
159
+ "unit": "tokens/s",
160
+ "value": 485.74734815531303
161
+ },
162
+ "energy": null,
163
+ "efficiency": null
164
+ },
165
+ "decode": {
166
+ "memory": {
167
+ "unit": "MB",
168
+ "max_ram": 2104.762368,
169
+ "max_global_vram": 6060.244992,
170
+ "max_process_vram": 0.0,
171
+ "max_reserved": 5381.292032,
172
+ "max_allocated": 5282.647552
173
+ },
174
+ "latency": {
175
+ "unit": "s",
176
+ "count": 2,
177
+ "total": 3.1293699951171874,
178
+ "mean": 1.5646849975585937,
179
+ "stdev": 0.0006091918945312358,
180
+ "p50": 1.5646849975585937,
181
+ "p90": 1.5651723510742186,
182
+ "p95": 1.5652332702636718,
183
+ "p99": 1.5652820056152343,
184
+ "values": [
185
+ 1.5640758056640625,
186
+ 1.565294189453125
187
+ ]
188
+ },
189
+ "throughput": {
190
+ "unit": "tokens/s",
191
+ "value": 81.1664968975611
192
+ },
193
+ "energy": null,
194
+ "efficiency": null
195
+ },
196
+ "per_token": {
197
+ "memory": null,
198
+ "latency": {
199
+ "unit": "s",
200
+ "count": 254,
201
+ "total": 3.128971271514895,
202
+ "mean": 0.012318784533523199,
203
+ "stdev": 7.379232209852449e-05,
204
+ "p50": 0.012302335739135742,
205
+ "p90": 0.01234227180480957,
206
+ "p95": 0.012369970989227295,
207
+ "p99": 0.012658063535690307,
208
+ "values": [
209
+ 0.012331007957458496,
210
+ 0.012318752288818359,
211
+ 0.012306400299072266,
212
+ 0.012288000106811523,
213
+ 0.01229312038421631,
214
+ 0.01229312038421631,
215
+ 0.012302335739135742,
216
+ 0.012309503555297852,
217
+ 0.012286975860595703,
218
+ 0.012299263954162597,
219
+ 0.012310527801513671,
220
+ 0.012286975860595703,
221
+ 0.012308480262756348,
222
+ 0.012296192169189453,
223
+ 0.01236684799194336,
224
+ 0.012329983711242675,
225
+ 0.012291071891784668,
226
+ 0.012297216415405274,
227
+ 0.012297216415405274,
228
+ 0.012302335739135742,
229
+ 0.012304384231567383,
230
+ 0.012291071891784668,
231
+ 0.012291071891784668,
232
+ 0.012298239707946777,
233
+ 0.012301312446594239,
234
+ 0.012295167922973632,
235
+ 0.012292096138000488,
236
+ 0.012306431770324706,
237
+ 0.012344320297241211,
238
+ 0.012291071891784668,
239
+ 0.012296192169189453,
240
+ 0.012297216415405274,
241
+ 0.012289024353027344,
242
+ 0.012309503555297852,
243
+ 0.012292096138000488,
244
+ 0.012311552047729492,
245
+ 0.012301312446594239,
246
+ 0.012309503555297852,
247
+ 0.012296192169189453,
248
+ 0.012298239707946777,
249
+ 0.012371968269348145,
250
+ 0.012352512359619141,
251
+ 0.012316672325134278,
252
+ 0.013006848335266113,
253
+ 0.01233407974243164,
254
+ 0.012302335739135742,
255
+ 0.012306431770324706,
256
+ 0.012315648078918457,
257
+ 0.012294143676757812,
258
+ 0.012309503555297852,
259
+ 0.012292096138000488,
260
+ 0.012318719863891601,
261
+ 0.012298239707946777,
262
+ 0.012286975860595703,
263
+ 0.012297216415405274,
264
+ 0.012296192169189453,
265
+ 0.012304384231567383,
266
+ 0.012314623832702636,
267
+ 0.01229312038421631,
268
+ 0.012315648078918457,
269
+ 0.012294143676757812,
270
+ 0.012411904335021973,
271
+ 0.01232588768005371,
272
+ 0.012311552047729492,
273
+ 0.012333056449890138,
274
+ 0.012310527801513671,
275
+ 0.012310527801513671,
276
+ 0.01233510398864746,
277
+ 0.012312576293945313,
278
+ 0.012310527801513671,
279
+ 0.012321791648864745,
280
+ 0.012288000106811523,
281
+ 0.012299263954162597,
282
+ 0.012308480262756348,
283
+ 0.012299263954162597,
284
+ 0.012301312446594239,
285
+ 0.012352512359619141,
286
+ 0.012320768356323243,
287
+ 0.012292096138000488,
288
+ 0.012298239707946777,
289
+ 0.012296192169189453,
290
+ 0.012305407524108887,
291
+ 0.012311552047729492,
292
+ 0.012299263954162597,
293
+ 0.012313599586486817,
294
+ 0.012295167922973632,
295
+ 0.012309503555297852,
296
+ 0.012292096138000488,
297
+ 0.012299263954162597,
298
+ 0.012354559898376465,
299
+ 0.012324864387512208,
300
+ 0.012288000106811523,
301
+ 0.012301312446594239,
302
+ 0.012295167922973632,
303
+ 0.012296192169189453,
304
+ 0.012295167922973632,
305
+ 0.012307456016540527,
306
+ 0.012280832290649414,
307
+ 0.012321791648864745,
308
+ 0.01229312038421631,
309
+ 0.012302335739135742,
310
+ 0.012295167922973632,
311
+ 0.012297216415405274,
312
+ 0.01248256015777588,
313
+ 0.012304384231567383,
314
+ 0.012302335739135742,
315
+ 0.012283904075622559,
316
+ 0.012288000106811523,
317
+ 0.012311552047729492,
318
+ 0.012298239707946777,
319
+ 0.01229312038421631,
320
+ 0.012289024353027344,
321
+ 0.012292096138000488,
322
+ 0.012288000106811523,
323
+ 0.012301312446594239,
324
+ 0.012306431770324706,
325
+ 0.012322815895080566,
326
+ 0.012315648078918457,
327
+ 0.012295167922973632,
328
+ 0.012329983711242675,
329
+ 0.012312576293945313,
330
+ 0.012327936172485352,
331
+ 0.012301312446594239,
332
+ 0.012296192169189453,
333
+ 0.012303359985351562,
334
+ 0.012302335739135742,
335
+ 0.012301312446594239,
336
+ 0.01257369613647461,
337
+ 0.012308480262756348,
338
+ 0.012314623832702636,
339
+ 0.012307456016540527,
340
+ 0.012319744110107422,
341
+ 0.012294143676757812,
342
+ 0.012298239707946777,
343
+ 0.012291071891784668,
344
+ 0.012304384231567383,
345
+ 0.01229312038421631,
346
+ 0.012297216415405274,
347
+ 0.012310527801513671,
348
+ 0.012400639533996583,
349
+ 0.012306431770324706,
350
+ 0.01234227180480957,
351
+ 0.012375040054321289,
352
+ 0.01234227180480957,
353
+ 0.012286975860595703,
354
+ 0.01229312038421631,
355
+ 0.012301312446594239,
356
+ 0.012296192169189453,
357
+ 0.012304384231567383,
358
+ 0.012298239707946777,
359
+ 0.012292096138000488,
360
+ 0.012301312446594239,
361
+ 0.012292096138000488,
362
+ 0.012295167922973632,
363
+ 0.012298239707946777,
364
+ 0.012311552047729492,
365
+ 0.01234329605102539,
366
+ 0.012304384231567383,
367
+ 0.012307456016540527,
368
+ 0.012300288200378418,
369
+ 0.012299263954162597,
370
+ 0.012308480262756348,
371
+ 0.012309503555297852,
372
+ 0.012312576293945313,
373
+ 0.012286975860595703,
374
+ 0.012297216415405274,
375
+ 0.012310527801513671,
376
+ 0.012320768356323243,
377
+ 0.012298239707946777,
378
+ 0.012297216415405274,
379
+ 0.012306431770324706,
380
+ 0.012289024353027344,
381
+ 0.012300288200378418,
382
+ 0.012289024353027344,
383
+ 0.012290047645568849,
384
+ 0.012322815895080566,
385
+ 0.012298239707946777,
386
+ 0.01229312038421631,
387
+ 0.012320768356323243,
388
+ 0.012289024353027344,
389
+ 0.012292096138000488,
390
+ 0.012299263954162597,
391
+ 0.012314623832702636,
392
+ 0.012346367835998535,
393
+ 0.012295167922973632,
394
+ 0.012289024353027344,
395
+ 0.012295167922973632,
396
+ 0.012291071891784668,
397
+ 0.012310527801513671,
398
+ 0.012319744110107422,
399
+ 0.012299263954162597,
400
+ 0.012678144454956054,
401
+ 0.012571647644042968,
402
+ 0.012640255928039551,
403
+ 0.012345343589782716,
404
+ 0.012300288200378418,
405
+ 0.012304384231567383,
406
+ 0.012304384231567383,
407
+ 0.012333056449890138,
408
+ 0.012307456016540527,
409
+ 0.012302335739135742,
410
+ 0.012315648078918457,
411
+ 0.012299263954162597,
412
+ 0.012301312446594239,
413
+ 0.012363776206970215,
414
+ 0.012332032203674317,
415
+ 0.012297216415405274,
416
+ 0.012289024353027344,
417
+ 0.012309503555297852,
418
+ 0.012310527801513671,
419
+ 0.012942336082458495,
420
+ 0.01234329605102539,
421
+ 0.012323840141296387,
422
+ 0.012314623832702636,
423
+ 0.012300288200378418,
424
+ 0.012301312446594239,
425
+ 0.012329983711242675,
426
+ 0.012326911926269531,
427
+ 0.012310527801513671,
428
+ 0.012314623832702636,
429
+ 0.012300288200378418,
430
+ 0.012309503555297852,
431
+ 0.0123505277633667,
432
+ 0.012393407821655274,
433
+ 0.012316672325134278,
434
+ 0.012309503555297852,
435
+ 0.01229312038421631,
436
+ 0.012304384231567383,
437
+ 0.012393471717834472,
438
+ 0.012368895530700684,
439
+ 0.012315648078918457,
440
+ 0.012302335739135742,
441
+ 0.012296192169189453,
442
+ 0.012311552047729492,
443
+ 0.01229312038421631,
444
+ 0.012307456016540527,
445
+ 0.012300352096557617,
446
+ 0.012302271842956543,
447
+ 0.012289024353027344,
448
+ 0.012295167922973632,
449
+ 0.012304384231567383,
450
+ 0.012321791648864745,
451
+ 0.012298239707946777,
452
+ 0.012298239707946777,
453
+ 0.012285951614379884,
454
+ 0.012294143676757812,
455
+ 0.012300288200378418,
456
+ 0.012305407524108887,
457
+ 0.01229312038421631,
458
+ 0.01229312038421631,
459
+ 0.012307456016540527,
460
+ 0.012292096138000488,
461
+ 0.012294143676757812,
462
+ 0.012298239707946777
463
+ ]
464
+ },
465
+ "throughput": {
466
+ "unit": "tokens/s",
467
+ "value": 81.1768399129551
468
+ },
469
+ "energy": null,
470
+ "efficiency": null
471
+ }
472
+ }
473
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark.log ADDED
File without changes
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_config.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "pytorch_generate",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0+cu121",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "library": "transformers",
9
+ "model_type": "gemma",
10
+ "model": "google/gemma-2b",
11
+ "processor": "google/gemma-2b",
12
+ "device": "cuda",
13
+ "device_ids": "0",
14
+ "seed": 42,
15
+ "inter_op_num_threads": null,
16
+ "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "no_weights": true,
20
+ "device_map": null,
21
+ "torch_dtype": "float16",
22
+ "eval_mode": true,
23
+ "to_bettertransformer": false,
24
+ "low_cpu_mem_usage": null,
25
+ "attn_implementation": null,
26
+ "cache_implementation": "static",
27
+ "autocast_enabled": false,
28
+ "autocast_dtype": null,
29
+ "torch_compile": true,
30
+ "torch_compile_target": "forward",
31
+ "torch_compile_config": {
32
+ "backend": "inductor",
33
+ "mode": "reduce-overhead",
34
+ "fullgraph": true
35
+ },
36
+ "quantization_scheme": null,
37
+ "quantization_config": {},
38
+ "deepspeed_inference": false,
39
+ "deepspeed_inference_config": {},
40
+ "peft_type": null,
41
+ "peft_config": {}
42
+ },
43
+ "scenario": {
44
+ "name": "inference",
45
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
46
+ "iterations": 2,
47
+ "duration": 0,
48
+ "warmup_runs": 10,
49
+ "input_shapes": {
50
+ "batch_size": 1,
51
+ "num_choices": 2,
52
+ "sequence_length": 7
53
+ },
54
+ "new_tokens": null,
55
+ "memory": true,
56
+ "latency": true,
57
+ "energy": false,
58
+ "forward_kwargs": {},
59
+ "generate_kwargs": {
60
+ "max_new_tokens": 128,
61
+ "min_new_tokens": 128,
62
+ "do_sample": false
63
+ },
64
+ "call_kwargs": {}
65
+ },
66
+ "launcher": {
67
+ "name": "process",
68
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
69
+ "device_isolation": true,
70
+ "device_isolation_action": "warn",
71
+ "numactl": false,
72
+ "numactl_kwargs": {},
73
+ "start_method": "spawn"
74
+ },
75
+ "environment": {
76
+ "cpu": " AMD EPYC 7R32",
77
+ "cpu_count": 16,
78
+ "cpu_ram_mb": 66697.261056,
79
+ "system": "Linux",
80
+ "machine": "x86_64",
81
+ "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29",
82
+ "processor": "x86_64",
83
+ "python_version": "3.8.10",
84
+ "gpu": [
85
+ "NVIDIA A10G"
86
+ ],
87
+ "gpu_count": 1,
88
+ "gpu_vram_mb": 24146608128,
89
+ "optimum_benchmark_version": "0.4.0",
90
+ "optimum_benchmark_commit": null,
91
+ "transformers_version": "4.45.0.dev0",
92
+ "transformers_commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
93
+ "accelerate_version": "0.35.0.dev0",
94
+ "accelerate_commit": null,
95
+ "diffusers_version": null,
96
+ "diffusers_commit": null,
97
+ "optimum_version": "1.23.0.dev0",
98
+ "optimum_commit": null,
99
+ "timm_version": "0.9.16",
100
+ "timm_commit": null,
101
+ "peft_version": "0.12.1.dev0",
102
+ "peft_commit": null
103
+ }
104
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/benchmark_report.json ADDED
@@ -0,0 +1,367 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "load": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1353.68704,
6
+ "max_global_vram": 6775.373824,
7
+ "max_process_vram": 0.0,
8
+ "max_reserved": 6138.363904,
9
+ "max_allocated": 6060.931072
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 1,
14
+ "total": 12.495619140625,
15
+ "mean": 12.495619140625,
16
+ "stdev": 0.0,
17
+ "p50": 12.495619140625,
18
+ "p90": 12.495619140625,
19
+ "p95": 12.495619140625,
20
+ "p99": 12.495619140625,
21
+ "values": [
22
+ 12.495619140625
23
+ ]
24
+ },
25
+ "throughput": null,
26
+ "energy": null,
27
+ "efficiency": null
28
+ },
29
+ "prefill": {
30
+ "memory": {
31
+ "unit": "MB",
32
+ "max_ram": 1931.149312,
33
+ "max_global_vram": 6796.345344,
34
+ "max_process_vram": 0.0,
35
+ "max_reserved": 6138.363904,
36
+ "max_allocated": 5278.596608
37
+ },
38
+ "latency": {
39
+ "unit": "s",
40
+ "count": 2,
41
+ "total": 0.02882156753540039,
42
+ "mean": 0.014410783767700195,
43
+ "stdev": 0.00048147201538085993,
44
+ "p50": 0.014410783767700195,
45
+ "p90": 0.014795961380004883,
46
+ "p95": 0.014844108581542968,
47
+ "p99": 0.014882626342773438,
48
+ "values": [
49
+ 0.013929311752319335,
50
+ 0.014892255783081055
51
+ ]
52
+ },
53
+ "throughput": {
54
+ "unit": "tokens/s",
55
+ "value": 485.74734815531303
56
+ },
57
+ "energy": null,
58
+ "efficiency": null
59
+ },
60
+ "decode": {
61
+ "memory": {
62
+ "unit": "MB",
63
+ "max_ram": 2104.762368,
64
+ "max_global_vram": 6060.244992,
65
+ "max_process_vram": 0.0,
66
+ "max_reserved": 5381.292032,
67
+ "max_allocated": 5282.647552
68
+ },
69
+ "latency": {
70
+ "unit": "s",
71
+ "count": 2,
72
+ "total": 3.1293699951171874,
73
+ "mean": 1.5646849975585937,
74
+ "stdev": 0.0006091918945312358,
75
+ "p50": 1.5646849975585937,
76
+ "p90": 1.5651723510742186,
77
+ "p95": 1.5652332702636718,
78
+ "p99": 1.5652820056152343,
79
+ "values": [
80
+ 1.5640758056640625,
81
+ 1.565294189453125
82
+ ]
83
+ },
84
+ "throughput": {
85
+ "unit": "tokens/s",
86
+ "value": 81.1664968975611
87
+ },
88
+ "energy": null,
89
+ "efficiency": null
90
+ },
91
+ "per_token": {
92
+ "memory": null,
93
+ "latency": {
94
+ "unit": "s",
95
+ "count": 254,
96
+ "total": 3.128971271514895,
97
+ "mean": 0.012318784533523199,
98
+ "stdev": 7.379232209852449e-05,
99
+ "p50": 0.012302335739135742,
100
+ "p90": 0.01234227180480957,
101
+ "p95": 0.012369970989227295,
102
+ "p99": 0.012658063535690307,
103
+ "values": [
104
+ 0.012331007957458496,
105
+ 0.012318752288818359,
106
+ 0.012306400299072266,
107
+ 0.012288000106811523,
108
+ 0.01229312038421631,
109
+ 0.01229312038421631,
110
+ 0.012302335739135742,
111
+ 0.012309503555297852,
112
+ 0.012286975860595703,
113
+ 0.012299263954162597,
114
+ 0.012310527801513671,
115
+ 0.012286975860595703,
116
+ 0.012308480262756348,
117
+ 0.012296192169189453,
118
+ 0.01236684799194336,
119
+ 0.012329983711242675,
120
+ 0.012291071891784668,
121
+ 0.012297216415405274,
122
+ 0.012297216415405274,
123
+ 0.012302335739135742,
124
+ 0.012304384231567383,
125
+ 0.012291071891784668,
126
+ 0.012291071891784668,
127
+ 0.012298239707946777,
128
+ 0.012301312446594239,
129
+ 0.012295167922973632,
130
+ 0.012292096138000488,
131
+ 0.012306431770324706,
132
+ 0.012344320297241211,
133
+ 0.012291071891784668,
134
+ 0.012296192169189453,
135
+ 0.012297216415405274,
136
+ 0.012289024353027344,
137
+ 0.012309503555297852,
138
+ 0.012292096138000488,
139
+ 0.012311552047729492,
140
+ 0.012301312446594239,
141
+ 0.012309503555297852,
142
+ 0.012296192169189453,
143
+ 0.012298239707946777,
144
+ 0.012371968269348145,
145
+ 0.012352512359619141,
146
+ 0.012316672325134278,
147
+ 0.013006848335266113,
148
+ 0.01233407974243164,
149
+ 0.012302335739135742,
150
+ 0.012306431770324706,
151
+ 0.012315648078918457,
152
+ 0.012294143676757812,
153
+ 0.012309503555297852,
154
+ 0.012292096138000488,
155
+ 0.012318719863891601,
156
+ 0.012298239707946777,
157
+ 0.012286975860595703,
158
+ 0.012297216415405274,
159
+ 0.012296192169189453,
160
+ 0.012304384231567383,
161
+ 0.012314623832702636,
162
+ 0.01229312038421631,
163
+ 0.012315648078918457,
164
+ 0.012294143676757812,
165
+ 0.012411904335021973,
166
+ 0.01232588768005371,
167
+ 0.012311552047729492,
168
+ 0.012333056449890138,
169
+ 0.012310527801513671,
170
+ 0.012310527801513671,
171
+ 0.01233510398864746,
172
+ 0.012312576293945313,
173
+ 0.012310527801513671,
174
+ 0.012321791648864745,
175
+ 0.012288000106811523,
176
+ 0.012299263954162597,
177
+ 0.012308480262756348,
178
+ 0.012299263954162597,
179
+ 0.012301312446594239,
180
+ 0.012352512359619141,
181
+ 0.012320768356323243,
182
+ 0.012292096138000488,
183
+ 0.012298239707946777,
184
+ 0.012296192169189453,
185
+ 0.012305407524108887,
186
+ 0.012311552047729492,
187
+ 0.012299263954162597,
188
+ 0.012313599586486817,
189
+ 0.012295167922973632,
190
+ 0.012309503555297852,
191
+ 0.012292096138000488,
192
+ 0.012299263954162597,
193
+ 0.012354559898376465,
194
+ 0.012324864387512208,
195
+ 0.012288000106811523,
196
+ 0.012301312446594239,
197
+ 0.012295167922973632,
198
+ 0.012296192169189453,
199
+ 0.012295167922973632,
200
+ 0.012307456016540527,
201
+ 0.012280832290649414,
202
+ 0.012321791648864745,
203
+ 0.01229312038421631,
204
+ 0.012302335739135742,
205
+ 0.012295167922973632,
206
+ 0.012297216415405274,
207
+ 0.01248256015777588,
208
+ 0.012304384231567383,
209
+ 0.012302335739135742,
210
+ 0.012283904075622559,
211
+ 0.012288000106811523,
212
+ 0.012311552047729492,
213
+ 0.012298239707946777,
214
+ 0.01229312038421631,
215
+ 0.012289024353027344,
216
+ 0.012292096138000488,
217
+ 0.012288000106811523,
218
+ 0.012301312446594239,
219
+ 0.012306431770324706,
220
+ 0.012322815895080566,
221
+ 0.012315648078918457,
222
+ 0.012295167922973632,
223
+ 0.012329983711242675,
224
+ 0.012312576293945313,
225
+ 0.012327936172485352,
226
+ 0.012301312446594239,
227
+ 0.012296192169189453,
228
+ 0.012303359985351562,
229
+ 0.012302335739135742,
230
+ 0.012301312446594239,
231
+ 0.01257369613647461,
232
+ 0.012308480262756348,
233
+ 0.012314623832702636,
234
+ 0.012307456016540527,
235
+ 0.012319744110107422,
236
+ 0.012294143676757812,
237
+ 0.012298239707946777,
238
+ 0.012291071891784668,
239
+ 0.012304384231567383,
240
+ 0.01229312038421631,
241
+ 0.012297216415405274,
242
+ 0.012310527801513671,
243
+ 0.012400639533996583,
244
+ 0.012306431770324706,
245
+ 0.01234227180480957,
246
+ 0.012375040054321289,
247
+ 0.01234227180480957,
248
+ 0.012286975860595703,
249
+ 0.01229312038421631,
250
+ 0.012301312446594239,
251
+ 0.012296192169189453,
252
+ 0.012304384231567383,
253
+ 0.012298239707946777,
254
+ 0.012292096138000488,
255
+ 0.012301312446594239,
256
+ 0.012292096138000488,
257
+ 0.012295167922973632,
258
+ 0.012298239707946777,
259
+ 0.012311552047729492,
260
+ 0.01234329605102539,
261
+ 0.012304384231567383,
262
+ 0.012307456016540527,
263
+ 0.012300288200378418,
264
+ 0.012299263954162597,
265
+ 0.012308480262756348,
266
+ 0.012309503555297852,
267
+ 0.012312576293945313,
268
+ 0.012286975860595703,
269
+ 0.012297216415405274,
270
+ 0.012310527801513671,
271
+ 0.012320768356323243,
272
+ 0.012298239707946777,
273
+ 0.012297216415405274,
274
+ 0.012306431770324706,
275
+ 0.012289024353027344,
276
+ 0.012300288200378418,
277
+ 0.012289024353027344,
278
+ 0.012290047645568849,
279
+ 0.012322815895080566,
280
+ 0.012298239707946777,
281
+ 0.01229312038421631,
282
+ 0.012320768356323243,
283
+ 0.012289024353027344,
284
+ 0.012292096138000488,
285
+ 0.012299263954162597,
286
+ 0.012314623832702636,
287
+ 0.012346367835998535,
288
+ 0.012295167922973632,
289
+ 0.012289024353027344,
290
+ 0.012295167922973632,
291
+ 0.012291071891784668,
292
+ 0.012310527801513671,
293
+ 0.012319744110107422,
294
+ 0.012299263954162597,
295
+ 0.012678144454956054,
296
+ 0.012571647644042968,
297
+ 0.012640255928039551,
298
+ 0.012345343589782716,
299
+ 0.012300288200378418,
300
+ 0.012304384231567383,
301
+ 0.012304384231567383,
302
+ 0.012333056449890138,
303
+ 0.012307456016540527,
304
+ 0.012302335739135742,
305
+ 0.012315648078918457,
306
+ 0.012299263954162597,
307
+ 0.012301312446594239,
308
+ 0.012363776206970215,
309
+ 0.012332032203674317,
310
+ 0.012297216415405274,
311
+ 0.012289024353027344,
312
+ 0.012309503555297852,
313
+ 0.012310527801513671,
314
+ 0.012942336082458495,
315
+ 0.01234329605102539,
316
+ 0.012323840141296387,
317
+ 0.012314623832702636,
318
+ 0.012300288200378418,
319
+ 0.012301312446594239,
320
+ 0.012329983711242675,
321
+ 0.012326911926269531,
322
+ 0.012310527801513671,
323
+ 0.012314623832702636,
324
+ 0.012300288200378418,
325
+ 0.012309503555297852,
326
+ 0.0123505277633667,
327
+ 0.012393407821655274,
328
+ 0.012316672325134278,
329
+ 0.012309503555297852,
330
+ 0.01229312038421631,
331
+ 0.012304384231567383,
332
+ 0.012393471717834472,
333
+ 0.012368895530700684,
334
+ 0.012315648078918457,
335
+ 0.012302335739135742,
336
+ 0.012296192169189453,
337
+ 0.012311552047729492,
338
+ 0.01229312038421631,
339
+ 0.012307456016540527,
340
+ 0.012300352096557617,
341
+ 0.012302271842956543,
342
+ 0.012289024353027344,
343
+ 0.012295167922973632,
344
+ 0.012304384231567383,
345
+ 0.012321791648864745,
346
+ 0.012298239707946777,
347
+ 0.012298239707946777,
348
+ 0.012285951614379884,
349
+ 0.012294143676757812,
350
+ 0.012300288200378418,
351
+ 0.012305407524108887,
352
+ 0.01229312038421631,
353
+ 0.01229312038421631,
354
+ 0.012307456016540527,
355
+ 0.012292096138000488,
356
+ 0.012294143676757812,
357
+ 0.012298239707946777
358
+ ]
359
+ },
360
+ "throughput": {
361
+ "unit": "tokens/s",
362
+ "value": 81.1768399129551
363
+ },
364
+ "energy": null,
365
+ "efficiency": null
366
+ }
367
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/backend.cache_implementation=static,backend.model=google/gemma-2b,backend.torch_compile=True/summary.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "google/gemma-2b",
3
+ "commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
4
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.014410783767700195,
7
+ "prefill.throughput.value": 485.74734815531303,
8
+ "decode.latency.mean": 1.5646849975585937,
9
+ "decode.throughput.value": 81.1664968975611,
10
+ "per_token.latency.mean": 0.012318784533523199,
11
+ "per_token.throughput.value": 81.1768399129551
12
+ }
13
+ }
2024-09-16/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/multirun.yaml ADDED
@@ -0,0 +1,234 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
4
+ sweep:
5
+ dir: _benchmark/tmpwgutcz5j/commit=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
6
+ subdir: ${hydra.job.override_dirname}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ root:
71
+ level: ERROR
72
+ disable_existing_loggers: true
73
+ job_logging:
74
+ version: 1
75
+ root:
76
+ level: ERROR
77
+ disable_existing_loggers: true
78
+ env: {}
79
+ mode: MULTIRUN
80
+ searchpath: []
81
+ callbacks: {}
82
+ output_subdir: .hydra
83
+ overrides:
84
+ hydra:
85
+ - hydra/job_logging=disabled
86
+ - hydra/hydra_logging=disabled
87
+ - hydra.sweep.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462
88
+ - hydra.run.dir=_benchmark/tmpwgutcz5j/commit\=8bd2b1e8c23234cd607ca8d63f53c1edfea27462/${hydra.job.override_dirname}
89
+ - hydra.mode=MULTIRUN
90
+ task:
91
+ - backend.model=google/gemma-2b
92
+ - backend.cache_implementation=null,static
93
+ - backend.torch_compile=false,true
94
+ job:
95
+ name: cli
96
+ chdir: true
97
+ override_dirname: backend.cache_implementation=null,static,backend.model=google/gemma-2b,backend.torch_compile=false,true
98
+ id: ???
99
+ num: ???
100
+ config_name: generation
101
+ env_set:
102
+ OVERRIDE_BENCHMARKS: '1'
103
+ LOG_LEVEL: WARN
104
+ env_copy: []
105
+ config:
106
+ override_dirname:
107
+ kv_sep: '='
108
+ item_sep: ','
109
+ exclude_keys: []
110
+ runtime:
111
+ version: 1.3.2
112
+ version_base: '1.3'
113
+ cwd: /transformers
114
+ config_sources:
115
+ - path: hydra.conf
116
+ schema: pkg
117
+ provider: hydra
118
+ - path: optimum_benchmark
119
+ schema: pkg
120
+ provider: main
121
+ - path: /transformers/benchmark/config
122
+ schema: file
123
+ provider: command-line
124
+ - path: ''
125
+ schema: structured
126
+ provider: schema
127
+ output_dir: ???
128
+ choices:
129
+ backend: pytorch
130
+ launcher: process
131
+ scenario: inference
132
+ hydra/env: default
133
+ hydra/callbacks: null
134
+ hydra/job_logging: disabled
135
+ hydra/hydra_logging: disabled
136
+ hydra/hydra_help: default
137
+ hydra/help: default
138
+ hydra/sweeper: basic
139
+ hydra/launcher: basic
140
+ hydra/output: default
141
+ verbose: false
142
+ name: pytorch_generate
143
+ backend:
144
+ name: pytorch
145
+ version: 2.4.0+cu121
146
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
147
+ task: null
148
+ library: null
149
+ model_type: null
150
+ model: google/gemma-2b
151
+ processor: null
152
+ device: cuda
153
+ device_ids: '0'
154
+ seed: 42
155
+ inter_op_num_threads: null
156
+ intra_op_num_threads: null
157
+ model_kwargs: {}
158
+ processor_kwargs: {}
159
+ no_weights: true
160
+ device_map: null
161
+ torch_dtype: float16
162
+ eval_mode: true
163
+ to_bettertransformer: false
164
+ low_cpu_mem_usage: null
165
+ attn_implementation: null
166
+ cache_implementation: static
167
+ autocast_enabled: false
168
+ autocast_dtype: null
169
+ torch_compile: true
170
+ torch_compile_target: forward
171
+ torch_compile_config:
172
+ backend: inductor
173
+ mode: reduce-overhead
174
+ fullgraph: true
175
+ quantization_scheme: null
176
+ quantization_config: {}
177
+ deepspeed_inference: false
178
+ deepspeed_inference_config: {}
179
+ peft_type: null
180
+ peft_config: {}
181
+ scenario:
182
+ name: inference
183
+ _target_: optimum_benchmark.scenarios.inference.scenario.InferenceScenario
184
+ iterations: 2
185
+ duration: 0
186
+ warmup_runs: 10
187
+ input_shapes:
188
+ batch_size: 1
189
+ sequence_length: 7
190
+ new_tokens: null
191
+ memory: true
192
+ latency: true
193
+ energy: false
194
+ forward_kwargs: {}
195
+ generate_kwargs:
196
+ max_new_tokens: 128
197
+ min_new_tokens: 128
198
+ do_sample: false
199
+ call_kwargs: {}
200
+ launcher:
201
+ name: process
202
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
203
+ device_isolation: true
204
+ device_isolation_action: warn
205
+ numactl: false
206
+ numactl_kwargs: {}
207
+ start_method: spawn
208
+ environment:
209
+ cpu: ' AMD EPYC 7R32'
210
+ cpu_count: 16
211
+ cpu_ram_mb: 66697.261056
212
+ system: Linux
213
+ machine: x86_64
214
+ platform: Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.29
215
+ processor: x86_64
216
+ python_version: 3.8.10
217
+ gpu:
218
+ - NVIDIA A10G
219
+ gpu_count: 1
220
+ gpu_vram_mb: 24146608128
221
+ optimum_benchmark_version: 0.4.0
222
+ optimum_benchmark_commit: null
223
+ transformers_version: 4.45.0.dev0
224
+ transformers_commit: 8bd2b1e8c23234cd607ca8d63f53c1edfea27462
225
+ accelerate_version: 0.35.0.dev0
226
+ accelerate_commit: null
227
+ diffusers_version: null
228
+ diffusers_commit: null
229
+ optimum_version: 1.23.0.dev0
230
+ optimum_commit: null
231
+ timm_version: 0.9.16
232
+ timm_commit: null
233
+ peft_version: 0.12.1.dev0
234
+ peft_commit: null
2024-09-16/summaries.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "model": "google/gemma-2b",
4
+ "commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
5
+ "config": "backend.cache_implementation=null,backend.torch_compile=False",
6
+ "metrics": {
7
+ "prefill.latency.mean": 0.02127888011932373,
8
+ "prefill.throughput.value": 328.964680506996,
9
+ "decode.latency.mean": 2.5476180419921874,
10
+ "decode.throughput.value": 49.85048696730397,
11
+ "per_token.latency.mean": 0.020058720731359767,
12
+ "per_token.throughput.value": 49.85362792536424
13
+ }
14
+ },
15
+ {
16
+ "model": "google/gemma-2b",
17
+ "commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
18
+ "config": "backend.cache_implementation=static,backend.torch_compile=False",
19
+ "metrics": {
20
+ "prefill.latency.mean": 0.022850063323974608,
21
+ "prefill.throughput.value": 306.3448840710871,
22
+ "decode.latency.mean": 2.6918625488281247,
23
+ "decode.throughput.value": 47.17922913831101,
24
+ "per_token.latency.mean": 0.02119472374503068,
25
+ "per_token.throughput.value": 47.18155386358646
26
+ }
27
+ },
28
+ {
29
+ "model": "google/gemma-2b",
30
+ "commit": "8bd2b1e8c23234cd607ca8d63f53c1edfea27462",
31
+ "config": "backend.cache_implementation=static,backend.torch_compile=True",
32
+ "metrics": {
33
+ "prefill.latency.mean": 0.014410783767700195,
34
+ "prefill.throughput.value": 485.74734815531303,
35
+ "decode.latency.mean": 1.5646849975585937,
36
+ "decode.throughput.value": 81.1664968975611,
37
+ "per_token.latency.mean": 0.012318784533523199,
38
+ "per_token.throughput.value": 81.1768399129551
39
+ }
40
+ }
41
+ ]
2024-09-16/summary.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "google/gemma-2b": {
3
+ "backend.cache_implementation=null,backend.torch_compile=False": {
4
+ "8bd2b1e8c23234cd607ca8d63f53c1edfea27462": {
5
+ "metrics": {
6
+ "prefill.latency.mean": 0.02127888011932373,
7
+ "prefill.throughput.value": 328.964680506996,
8
+ "decode.latency.mean": 2.5476180419921874,
9
+ "decode.throughput.value": 49.85048696730397,
10
+ "per_token.latency.mean": 0.020058720731359767,
11
+ "per_token.throughput.value": 49.85362792536424
12
+ }
13
+ }
14
+ },
15
+ "backend.cache_implementation=static,backend.torch_compile=False": {
16
+ "8bd2b1e8c23234cd607ca8d63f53c1edfea27462": {
17
+ "metrics": {
18
+ "prefill.latency.mean": 0.022850063323974608,
19
+ "prefill.throughput.value": 306.3448840710871,
20
+ "decode.latency.mean": 2.6918625488281247,
21
+ "decode.throughput.value": 47.17922913831101,
22
+ "per_token.latency.mean": 0.02119472374503068,
23
+ "per_token.throughput.value": 47.18155386358646
24
+ }
25
+ }
26
+ },
27
+ "backend.cache_implementation=static,backend.torch_compile=True": {
28
+ "8bd2b1e8c23234cd607ca8d63f53c1edfea27462": {
29
+ "metrics": {
30
+ "prefill.latency.mean": 0.014410783767700195,
31
+ "prefill.throughput.value": 485.74734815531303,
32
+ "decode.latency.mean": 1.5646849975585937,
33
+ "decode.throughput.value": 81.1664968975611,
34
+ "per_token.latency.mean": 0.012318784533523199,
35
+ "per_token.throughput.value": 81.1768399129551
36
+ }
37
+ }
38
+ }
39
+ }
40
+ }