sasha HF Staff commited on
Commit
27a6740
·
verified ·
1 Parent(s): 28c243a

Upload folder using huggingface_hub

Browse files
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/.hydra/config.yaml ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ backend:
2
+ name: pytorch
3
+ version: 2.4.0
4
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
+ task: text-generation
6
+ model: EleutherAI/pythia-1.4b
7
+ processor: EleutherAI/pythia-1.4b
8
+ library: null
9
+ device: cuda
10
+ device_ids: '0'
11
+ seed: 42
12
+ inter_op_num_threads: null
13
+ intra_op_num_threads: null
14
+ hub_kwargs: {}
15
+ no_weights: true
16
+ device_map: null
17
+ torch_dtype: null
18
+ amp_autocast: false
19
+ amp_dtype: null
20
+ eval_mode: true
21
+ to_bettertransformer: false
22
+ low_cpu_mem_usage: null
23
+ attn_implementation: null
24
+ cache_implementation: null
25
+ torch_compile: false
26
+ torch_compile_config: {}
27
+ quantization_scheme: null
28
+ quantization_config: {}
29
+ deepspeed_inference: false
30
+ deepspeed_inference_config: {}
31
+ peft_type: null
32
+ peft_config: {}
33
+ launcher:
34
+ name: process
35
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
36
+ device_isolation: false
37
+ device_isolation_action: warn
38
+ start_method: spawn
39
+ benchmark:
40
+ name: energy_star
41
+ _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark
42
+ dataset_name: EnergyStarAI/text_generation
43
+ dataset_config: ''
44
+ dataset_split: train
45
+ num_samples: 1000
46
+ input_shapes:
47
+ batch_size: 1
48
+ text_column_name: text
49
+ truncation: true
50
+ max_length: -1
51
+ dataset_prefix1: ''
52
+ dataset_prefix2: ''
53
+ t5_task: ''
54
+ image_column_name: image
55
+ resize: false
56
+ question_column_name: question
57
+ context_column_name: context
58
+ sentence1_column_name: sentence1
59
+ sentence2_column_name: sentence2
60
+ audio_column_name: audio
61
+ iterations: 10
62
+ warmup_runs: 10
63
+ energy: true
64
+ forward_kwargs: {}
65
+ generate_kwargs:
66
+ max_new_tokens: 10
67
+ min_new_tokens: 10
68
+ call_kwargs: {}
69
+ experiment_name: text_generation
70
+ environment:
71
+ cpu: ' AMD EPYC 7R32'
72
+ cpu_count: 48
73
+ cpu_ram_mb: 200472.73984
74
+ system: Linux
75
+ machine: x86_64
76
+ platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
77
+ processor: x86_64
78
+ python_version: 3.9.20
79
+ gpu:
80
+ - NVIDIA A10G
81
+ gpu_count: 1
82
+ gpu_vram_mb: 24146608128
83
+ optimum_benchmark_version: 0.2.0
84
+ optimum_benchmark_commit: null
85
+ transformers_version: 4.44.0
86
+ transformers_commit: null
87
+ accelerate_version: 0.33.0
88
+ accelerate_commit: null
89
+ diffusers_version: 0.30.0
90
+ diffusers_commit: null
91
+ optimum_version: null
92
+ optimum_commit: null
93
+ timm_version: null
94
+ timm_commit: null
95
+ peft_version: null
96
+ peft_commit: null
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/.hydra/hydra.yaml ADDED
@@ -0,0 +1,175 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: /runs/text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57
4
+ sweep:
5
+ dir: sweeps/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ colorlog:
72
+ (): colorlog.ColoredFormatter
73
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
74
+ handlers:
75
+ console:
76
+ class: logging.StreamHandler
77
+ formatter: colorlog
78
+ stream: ext://sys.stdout
79
+ root:
80
+ level: INFO
81
+ handlers:
82
+ - console
83
+ disable_existing_loggers: false
84
+ job_logging:
85
+ version: 1
86
+ formatters:
87
+ simple:
88
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
89
+ colorlog:
90
+ (): colorlog.ColoredFormatter
91
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
92
+ - %(message)s'
93
+ log_colors:
94
+ DEBUG: purple
95
+ INFO: green
96
+ WARNING: yellow
97
+ ERROR: red
98
+ CRITICAL: red
99
+ handlers:
100
+ console:
101
+ class: logging.StreamHandler
102
+ formatter: colorlog
103
+ stream: ext://sys.stdout
104
+ file:
105
+ class: logging.FileHandler
106
+ formatter: simple
107
+ filename: ${hydra.job.name}.log
108
+ root:
109
+ level: INFO
110
+ handlers:
111
+ - console
112
+ - file
113
+ disable_existing_loggers: false
114
+ env: {}
115
+ mode: RUN
116
+ searchpath: []
117
+ callbacks: {}
118
+ output_subdir: .hydra
119
+ overrides:
120
+ hydra:
121
+ - hydra.run.dir=/runs/text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57
122
+ - hydra.mode=RUN
123
+ task:
124
+ - backend.model=EleutherAI/pythia-1.4b
125
+ - backend.processor=EleutherAI/pythia-1.4b
126
+ job:
127
+ name: cli
128
+ chdir: true
129
+ override_dirname: backend.model=EleutherAI/pythia-1.4b,backend.processor=EleutherAI/pythia-1.4b
130
+ id: ???
131
+ num: ???
132
+ config_name: text_generation
133
+ env_set:
134
+ OVERRIDE_BENCHMARKS: '1'
135
+ env_copy: []
136
+ config:
137
+ override_dirname:
138
+ kv_sep: '='
139
+ item_sep: ','
140
+ exclude_keys: []
141
+ runtime:
142
+ version: 1.3.2
143
+ version_base: '1.3'
144
+ cwd: /
145
+ config_sources:
146
+ - path: hydra.conf
147
+ schema: pkg
148
+ provider: hydra
149
+ - path: optimum_benchmark
150
+ schema: pkg
151
+ provider: main
152
+ - path: hydra_plugins.hydra_colorlog.conf
153
+ schema: pkg
154
+ provider: hydra-colorlog
155
+ - path: /optimum-benchmark/examples/energy_star
156
+ schema: file
157
+ provider: command-line
158
+ - path: ''
159
+ schema: structured
160
+ provider: schema
161
+ output_dir: /runs/text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57
162
+ choices:
163
+ benchmark: energy_star
164
+ launcher: process
165
+ backend: pytorch
166
+ hydra/env: default
167
+ hydra/callbacks: null
168
+ hydra/job_logging: colorlog
169
+ hydra/hydra_logging: colorlog
170
+ hydra/hydra_help: default
171
+ hydra/help: default
172
+ hydra/sweeper: basic
173
+ hydra/launcher: basic
174
+ hydra/output: default
175
+ verbose: false
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/.hydra/overrides.yaml ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ - backend.model=EleutherAI/pythia-1.4b
2
+ - backend.processor=EleutherAI/pythia-1.4b
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/benchmark_report.json ADDED
@@ -0,0 +1,203 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "prefill": {
3
+ "memory": null,
4
+ "latency": null,
5
+ "throughput": null,
6
+ "energy": {
7
+ "unit": "kWh",
8
+ "cpu": 0.0006784651691832389,
9
+ "ram": 6.872767178977215e-06,
10
+ "gpu": 0.004670849625565543,
11
+ "total": 0.00535618756192776
12
+ },
13
+ "efficiency": {
14
+ "unit": "tokens/kWh",
15
+ "value": 56280142.64151448
16
+ },
17
+ "measures": [
18
+ {
19
+ "unit": "kWh",
20
+ "cpu": 0.0007471547884041582,
21
+ "ram": 7.565518406011505e-06,
22
+ "gpu": 0.005128198824777996,
23
+ "total": 0.005882919131588165
24
+ },
25
+ {
26
+ "unit": "kWh",
27
+ "cpu": 0.0007498932815120901,
28
+ "ram": 7.596245961500314e-06,
29
+ "gpu": 0.005162669130131636,
30
+ "total": 0.005920158657605227
31
+ },
32
+ {
33
+ "unit": "kWh",
34
+ "cpu": 0.0007506841969611641,
35
+ "ram": 7.6041352101112216e-06,
36
+ "gpu": 0.005159375238607655,
37
+ "total": 0.005917663570778931
38
+ },
39
+ {
40
+ "unit": "kWh",
41
+ "cpu": 0.0007552490596986899,
42
+ "ram": 7.65116218203134e-06,
43
+ "gpu": 0.00519234248720446,
44
+ "total": 0.005955242709085182
45
+ },
46
+ {
47
+ "unit": "kWh",
48
+ "cpu": 0.0007571123219730605,
49
+ "ram": 7.669590319577743e-06,
50
+ "gpu": 0.00525252309090396,
51
+ "total": 0.006017305003196596
52
+ },
53
+ {
54
+ "unit": "kWh",
55
+ "cpu": 0.0007562589633509356,
56
+ "ram": 7.661224940613736e-06,
57
+ "gpu": 0.005222808067132156,
58
+ "total": 0.005986728255423707
59
+ },
60
+ {
61
+ "unit": "kWh",
62
+ "cpu": 0.0,
63
+ "ram": 0.0,
64
+ "gpu": 0.0,
65
+ "total": 0.0
66
+ },
67
+ {
68
+ "unit": "kWh",
69
+ "cpu": 0.0007560068380162398,
70
+ "ram": 7.658837285322218e-06,
71
+ "gpu": 0.0052021186061357305,
72
+ "total": 0.005965784281437292
73
+ },
74
+ {
75
+ "unit": "kWh",
76
+ "cpu": 0.0007560777030474536,
77
+ "ram": 7.659780890775818e-06,
78
+ "gpu": 0.005200013604452103,
79
+ "total": 0.005963751088390333
80
+ },
81
+ {
82
+ "unit": "kWh",
83
+ "cpu": 0.000756214538868598,
84
+ "ram": 7.661176593828245e-06,
85
+ "gpu": 0.005188447206309732,
86
+ "total": 0.0059523229217721615
87
+ }
88
+ ]
89
+ },
90
+ "decode": {
91
+ "memory": null,
92
+ "latency": null,
93
+ "throughput": null,
94
+ "energy": {
95
+ "unit": "kWh",
96
+ "cpu": 0.0015004224549616597,
97
+ "ram": 1.5204797701429561e-05,
98
+ "gpu": 0.00586308749602189,
99
+ "total": 0.00737871474868498
100
+ },
101
+ "efficiency": {
102
+ "unit": "tokens/kWh",
103
+ "value": 1219724.6141821598
104
+ },
105
+ "measures": [
106
+ {
107
+ "unit": "kWh",
108
+ "cpu": 0.0016782613314461846,
109
+ "ram": 1.7009955692935013e-05,
110
+ "gpu": 0.006698563692180137,
111
+ "total": 0.008393834979319253
112
+ },
113
+ {
114
+ "unit": "kWh",
115
+ "cpu": 0.0016719845362669326,
116
+ "ram": 1.6942875567914424e-05,
117
+ "gpu": 0.0065212854948022425,
118
+ "total": 0.008210212906637095
119
+ },
120
+ {
121
+ "unit": "kWh",
122
+ "cpu": 0.0016724901459935718,
123
+ "ram": 1.6948727439907015e-05,
124
+ "gpu": 0.006527124666140693,
125
+ "total": 0.00821656353957417
126
+ },
127
+ {
128
+ "unit": "kWh",
129
+ "cpu": 0.0016649095729317186,
130
+ "ram": 1.6871157784986704e-05,
131
+ "gpu": 0.0064939879729633,
132
+ "total": 0.008175768703680013
133
+ },
134
+ {
135
+ "unit": "kWh",
136
+ "cpu": -0.0007571123219730605,
137
+ "ram": -7.669590319577743e-06,
138
+ "gpu": -0.00525252309090396,
139
+ "total": -0.006017305003196596
140
+ },
141
+ {
142
+ "unit": "kWh",
143
+ "cpu": 0.0016639060387660326,
144
+ "ram": 1.6861220430135516e-05,
145
+ "gpu": 0.00647061517648817,
146
+ "total": 0.008151382435684335
147
+ },
148
+ {
149
+ "unit": "kWh",
150
+ "cpu": 0.002422301665296781,
151
+ "ram": 2.4544091842681722e-05,
152
+ "gpu": 0.011690915463836049,
153
+ "total": 0.014137761220975514
154
+ },
155
+ {
156
+ "unit": "kWh",
157
+ "cpu": 0.0016633325706456384,
158
+ "ram": 1.6854834230008524e-05,
159
+ "gpu": 0.006496602697278142,
160
+ "total": 0.008176790102153775
161
+ },
162
+ {
163
+ "unit": "kWh",
164
+ "cpu": 0.0016610744073210513,
165
+ "ram": 1.6832224697802917e-05,
166
+ "gpu": 0.0064857387996974936,
167
+ "total": 0.008163645431716333
168
+ },
169
+ {
170
+ "unit": "kWh",
171
+ "cpu": 0.0016630766029217424,
172
+ "ram": 1.685247964750152e-05,
173
+ "gpu": 0.006498564087736636,
174
+ "total": 0.008178493170305896
175
+ }
176
+ ]
177
+ },
178
+ "per_token": {
179
+ "memory": null,
180
+ "latency": null,
181
+ "throughput": null,
182
+ "energy": null,
183
+ "efficiency": null,
184
+ "measures": null
185
+ },
186
+ "preprocess": {
187
+ "memory": null,
188
+ "latency": null,
189
+ "throughput": null,
190
+ "energy": {
191
+ "unit": "kWh",
192
+ "cpu": 1.1119852669475626e-05,
193
+ "ram": 8.376988347301967e-08,
194
+ "gpu": 1.861168155592452e-05,
195
+ "total": 2.9815304108873166e-05
196
+ },
197
+ "efficiency": {
198
+ "unit": "samples/kWh",
199
+ "value": 33539822.2452608
200
+ },
201
+ "measures": null
202
+ }
203
+ }
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/cli.log ADDED
@@ -0,0 +1,188 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2024-10-25 13:07:00,323][launcher][INFO] - ََAllocating process launcher
2
+ [2024-10-25 13:07:00,323][process][INFO] - + Setting multiprocessing start method to spawn.
3
+ [2024-10-25 13:07:00,335][process][INFO] - + Launched benchmark in isolated process 351.
4
+ [PROC-0][2024-10-25 13:07:02,873][datasets][INFO] - PyTorch version 2.4.0 available.
5
+ [PROC-0][2024-10-25 13:07:03,794][backend][INFO] - َAllocating pytorch backend
6
+ [PROC-0][2024-10-25 13:07:03,794][backend][INFO] - + Setting random seed to 42
7
+ [PROC-0][2024-10-25 13:07:04,645][pytorch][INFO] - + Using AutoModel class AutoModelForCausalLM
8
+ [PROC-0][2024-10-25 13:07:04,645][pytorch][INFO] - + Creating backend temporary directory
9
+ [PROC-0][2024-10-25 13:07:04,645][pytorch][INFO] - + Loading model with random weights
10
+ [PROC-0][2024-10-25 13:07:04,646][pytorch][INFO] - + Creating no weights model
11
+ [PROC-0][2024-10-25 13:07:04,646][pytorch][INFO] - + Creating no weights model directory
12
+ [PROC-0][2024-10-25 13:07:04,646][pytorch][INFO] - + Creating no weights model state dict
13
+ [PROC-0][2024-10-25 13:07:04,666][pytorch][INFO] - + Saving no weights model safetensors
14
+ [PROC-0][2024-10-25 13:07:04,666][pytorch][INFO] - + Saving no weights model pretrained config
15
+ [PROC-0][2024-10-25 13:07:04,667][pytorch][INFO] - + Loading no weights AutoModel
16
+ [PROC-0][2024-10-25 13:07:04,667][pytorch][INFO] - + Loading model directly on device: cuda
17
+ [PROC-0][2024-10-25 13:07:05,038][pytorch][INFO] - + Turning on model's eval mode
18
+ [PROC-0][2024-10-25 13:07:05,045][benchmark][INFO] - Allocating energy_star benchmark
19
+ [PROC-0][2024-10-25 13:07:05,045][energy_star][INFO] - + Loading raw dataset
20
+ [PROC-0][2024-10-25 13:07:06,316][energy_star][INFO] - + Updating Text Generation kwargs with default values
21
+ [PROC-0][2024-10-25 13:07:06,317][energy_star][INFO] - + Initializing Text Generation report
22
+ [PROC-0][2024-10-25 13:07:06,317][energy][INFO] - + Tracking GPU energy on devices [0]
23
+ [PROC-0][2024-10-25 13:07:10,519][energy_star][INFO] - + Preprocessing dataset
24
+ [PROC-0][2024-10-25 13:07:11,462][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json
25
+ [PROC-0][2024-10-25 13:07:11,463][energy_star][INFO] - + Preparing backend for Inference
26
+ [PROC-0][2024-10-25 13:07:11,463][energy_star][INFO] - + Initialising dataloader
27
+ [PROC-0][2024-10-25 13:07:11,463][energy_star][INFO] - + Warming up backend for Inference
28
+ [PROC-0][2024-10-25 13:07:12,868][energy_star][INFO] - + Additional warmup for Text Generation
29
+ [PROC-0][2024-10-25 13:07:13,067][energy_star][INFO] - + Running Text Generation energy tracking for 10 iterations
30
+ [PROC-0][2024-10-25 13:07:13,067][energy_star][INFO] - + Prefill iteration 1/10
31
+ [PROC-0][2024-10-25 13:08:16,356][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
32
+ [PROC-0][2024-10-25 13:08:16,357][energy_star][INFO] - + Prefill iteration 2/10
33
+ [PROC-0][2024-10-25 13:09:19,878][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
34
+ [PROC-0][2024-10-25 13:09:19,878][energy_star][INFO] - + Prefill iteration 3/10
35
+ [PROC-0][2024-10-25 13:10:23,466][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
36
+ [PROC-0][2024-10-25 13:10:23,467][energy_star][INFO] - + Prefill iteration 4/10
37
+ [PROC-0][2024-10-25 13:11:27,441][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
38
+ [PROC-0][2024-10-25 13:11:27,442][energy_star][INFO] - + Prefill iteration 5/10
39
+ [PROC-0][2024-10-25 13:12:31,574][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
40
+ [PROC-0][2024-10-25 13:12:31,574][energy_star][INFO] - + Prefill iteration 6/10
41
+ [PROC-0][2024-10-25 13:13:35,635][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
42
+ [PROC-0][2024-10-25 13:13:35,635][energy_star][INFO] - + Prefill iteration 7/10
43
+ [PROC-0][2024-10-25 13:14:39,660][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
44
+ [PROC-0][2024-10-25 13:14:39,660][energy_star][INFO] - + Prefill iteration 8/10
45
+ [PROC-0][2024-10-25 13:15:43,699][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
46
+ [PROC-0][2024-10-25 13:15:43,700][energy_star][INFO] - + Prefill iteration 9/10
47
+ [PROC-0][2024-10-25 13:16:47,745][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
48
+ [PROC-0][2024-10-25 13:16:47,745][energy_star][INFO] - + Prefill iteration 10/10
49
+ [PROC-0][2024-10-25 13:17:51,801][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
50
+ [PROC-0][2024-10-25 13:17:51,802][energy_star][INFO] - + Decoding iteration 1/10
51
+ [PROC-0][2024-10-25 13:21:17,249][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
52
+ [PROC-0][2024-10-25 13:21:17,250][energy_star][INFO] - + Decoding iteration 2/10
53
+ [PROC-0][2024-10-25 13:24:42,398][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
54
+ [PROC-0][2024-10-25 13:24:42,398][energy_star][INFO] - + Decoding iteration 3/10
55
+ [PROC-0][2024-10-25 13:28:07,656][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
56
+ [PROC-0][2024-10-25 13:28:07,656][energy_star][INFO] - + Decoding iteration 4/10
57
+ [PROC-0][2024-10-25 13:31:32,659][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
58
+ [PROC-0][2024-10-25 13:31:32,659][energy_star][INFO] - + Decoding iteration 5/10
59
+ [PROC-0][2024-10-25 13:34:57,839][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
60
+ [PROC-0][2024-10-25 13:34:57,840][energy_star][INFO] - + Decoding iteration 6/10
61
+ [PROC-0][2024-10-25 13:38:22,843][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
62
+ [PROC-0][2024-10-25 13:38:22,843][energy_star][INFO] - + Decoding iteration 7/10
63
+ [PROC-0][2024-10-25 13:41:48,027][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
64
+ [PROC-0][2024-10-25 13:41:48,027][energy_star][INFO] - + Decoding iteration 8/10
65
+ [PROC-0][2024-10-25 13:45:12,960][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
66
+ [PROC-0][2024-10-25 13:45:12,960][energy_star][INFO] - + Decoding iteration 9/10
67
+ [PROC-0][2024-10-25 13:48:37,708][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
68
+ [PROC-0][2024-10-25 13:48:37,708][energy_star][INFO] - + Decoding iteration 10/10
69
+ [PROC-0][2024-10-25 13:52:02,637][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
70
+ [PROC-0][2024-10-25 13:52:02,638][energy][INFO] - + prefill energy consumption:
71
+ [PROC-0][2024-10-25 13:52:02,638][energy][INFO] - + CPU: 0.000678 (kWh)
72
+ [PROC-0][2024-10-25 13:52:02,638][energy][INFO] - + GPU: 0.004671 (kWh)
73
+ [PROC-0][2024-10-25 13:52:02,638][energy][INFO] - + RAM: 0.000007 (kWh)
74
+ [PROC-0][2024-10-25 13:52:02,638][energy][INFO] - + total: 0.005356 (kWh)
75
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + prefill_iteration_1 energy consumption:
76
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + CPU: 0.000747 (kWh)
77
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + GPU: 0.005128 (kWh)
78
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + RAM: 0.000008 (kWh)
79
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + total: 0.005883 (kWh)
80
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + prefill_iteration_2 energy consumption:
81
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + CPU: 0.000750 (kWh)
82
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + GPU: 0.005163 (kWh)
83
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + RAM: 0.000008 (kWh)
84
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + total: 0.005920 (kWh)
85
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + prefill_iteration_3 energy consumption:
86
+ [PROC-0][2024-10-25 13:52:02,639][energy][INFO] - + CPU: 0.000751 (kWh)
87
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + GPU: 0.005159 (kWh)
88
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + RAM: 0.000008 (kWh)
89
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + total: 0.005918 (kWh)
90
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + prefill_iteration_4 energy consumption:
91
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + CPU: 0.000755 (kWh)
92
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + GPU: 0.005192 (kWh)
93
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + RAM: 0.000008 (kWh)
94
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + total: 0.005955 (kWh)
95
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + prefill_iteration_5 energy consumption:
96
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + CPU: 0.000757 (kWh)
97
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + GPU: 0.005253 (kWh)
98
+ [PROC-0][2024-10-25 13:52:02,640][energy][INFO] - + RAM: 0.000008 (kWh)
99
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + total: 0.006017 (kWh)
100
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + prefill_iteration_6 energy consumption:
101
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + CPU: 0.000756 (kWh)
102
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + GPU: 0.005223 (kWh)
103
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + RAM: 0.000008 (kWh)
104
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + total: 0.005987 (kWh)
105
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + prefill_iteration_7 energy consumption:
106
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + CPU: 0.000000 (kWh)
107
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + GPU: 0.000000 (kWh)
108
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + RAM: 0.000000 (kWh)
109
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + total: 0.000000 (kWh)
110
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + prefill_iteration_8 energy consumption:
111
+ [PROC-0][2024-10-25 13:52:02,641][energy][INFO] - + CPU: 0.000756 (kWh)
112
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + GPU: 0.005202 (kWh)
113
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + RAM: 0.000008 (kWh)
114
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + total: 0.005966 (kWh)
115
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + prefill_iteration_9 energy consumption:
116
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + CPU: 0.000756 (kWh)
117
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + GPU: 0.005200 (kWh)
118
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + RAM: 0.000008 (kWh)
119
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + total: 0.005964 (kWh)
120
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + prefill_iteration_10 energy consumption:
121
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + CPU: 0.000756 (kWh)
122
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + GPU: 0.005188 (kWh)
123
+ [PROC-0][2024-10-25 13:52:02,642][energy][INFO] - + RAM: 0.000008 (kWh)
124
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + total: 0.005952 (kWh)
125
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + decode energy consumption:
126
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + CPU: 0.001500 (kWh)
127
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + GPU: 0.005863 (kWh)
128
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + RAM: 0.000015 (kWh)
129
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + total: 0.007379 (kWh)
130
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + decode_iteration_1 energy consumption:
131
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + CPU: 0.001678 (kWh)
132
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + GPU: 0.006699 (kWh)
133
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + RAM: 0.000017 (kWh)
134
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + total: 0.008394 (kWh)
135
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + decode_iteration_2 energy consumption:
136
+ [PROC-0][2024-10-25 13:52:02,643][energy][INFO] - + CPU: 0.001672 (kWh)
137
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + GPU: 0.006521 (kWh)
138
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + RAM: 0.000017 (kWh)
139
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + total: 0.008210 (kWh)
140
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + decode_iteration_3 energy consumption:
141
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + CPU: 0.001672 (kWh)
142
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + GPU: 0.006527 (kWh)
143
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + RAM: 0.000017 (kWh)
144
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + total: 0.008217 (kWh)
145
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + decode_iteration_4 energy consumption:
146
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + CPU: 0.001665 (kWh)
147
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + GPU: 0.006494 (kWh)
148
+ [PROC-0][2024-10-25 13:52:02,644][energy][INFO] - + RAM: 0.000017 (kWh)
149
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + total: 0.008176 (kWh)
150
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + decode_iteration_5 energy consumption:
151
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + CPU: -0.000757 (kWh)
152
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + GPU: -0.005253 (kWh)
153
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + RAM: -0.000008 (kWh)
154
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + total: -0.006017 (kWh)
155
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + decode_iteration_6 energy consumption:
156
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + CPU: 0.001664 (kWh)
157
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + GPU: 0.006471 (kWh)
158
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + RAM: 0.000017 (kWh)
159
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + total: 0.008151 (kWh)
160
+ [PROC-0][2024-10-25 13:52:02,645][energy][INFO] - + decode_iteration_7 energy consumption:
161
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + CPU: 0.002422 (kWh)
162
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + GPU: 0.011691 (kWh)
163
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + RAM: 0.000025 (kWh)
164
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + total: 0.014138 (kWh)
165
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + decode_iteration_8 energy consumption:
166
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + CPU: 0.001663 (kWh)
167
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + GPU: 0.006497 (kWh)
168
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + RAM: 0.000017 (kWh)
169
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + total: 0.008177 (kWh)
170
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + decode_iteration_9 energy consumption:
171
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + CPU: 0.001661 (kWh)
172
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + GPU: 0.006486 (kWh)
173
+ [PROC-0][2024-10-25 13:52:02,646][energy][INFO] - + RAM: 0.000017 (kWh)
174
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + total: 0.008164 (kWh)
175
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + decode_iteration_10 energy consumption:
176
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + CPU: 0.001663 (kWh)
177
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + GPU: 0.006499 (kWh)
178
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + RAM: 0.000017 (kWh)
179
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + total: 0.008178 (kWh)
180
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + preprocess energy consumption:
181
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + CPU: 0.000011 (kWh)
182
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + GPU: 0.000019 (kWh)
183
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + RAM: 0.000000 (kWh)
184
+ [PROC-0][2024-10-25 13:52:02,647][energy][INFO] - + total: 0.000030 (kWh)
185
+ [PROC-0][2024-10-25 13:52:02,648][energy][INFO] - + prefill energy efficiency: 56280142.641514 (tokens/kWh)
186
+ [PROC-0][2024-10-25 13:52:02,648][energy][INFO] - + decode energy efficiency: 1219724.614182 (tokens/kWh)
187
+ [PROC-0][2024-10-25 13:52:02,648][energy][INFO] - + preprocess energy efficiency: 33539822.245261 (samples/kWh)
188
+ [2024-10-25 13:52:03,342][datasets][INFO] - PyTorch version 2.4.0 available.
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/error.log ADDED
The diff for this file is too large to render. See raw diff
 
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/experiment_config.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_name": "text_generation",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "model": "EleutherAI/pythia-1.4b",
9
+ "processor": "EleutherAI/pythia-1.4b",
10
+ "library": "transformers",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "hub_kwargs": {
17
+ "revision": "main",
18
+ "force_download": false,
19
+ "local_files_only": false,
20
+ "trust_remote_code": true
21
+ },
22
+ "no_weights": true,
23
+ "device_map": null,
24
+ "torch_dtype": null,
25
+ "amp_autocast": false,
26
+ "amp_dtype": null,
27
+ "eval_mode": true,
28
+ "to_bettertransformer": false,
29
+ "low_cpu_mem_usage": null,
30
+ "attn_implementation": null,
31
+ "cache_implementation": null,
32
+ "torch_compile": false,
33
+ "torch_compile_config": {},
34
+ "quantization_scheme": null,
35
+ "quantization_config": {},
36
+ "deepspeed_inference": false,
37
+ "deepspeed_inference_config": {},
38
+ "peft_type": null,
39
+ "peft_config": {}
40
+ },
41
+ "launcher": {
42
+ "name": "process",
43
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
44
+ "device_isolation": false,
45
+ "device_isolation_action": "warn",
46
+ "start_method": "spawn"
47
+ },
48
+ "benchmark": {
49
+ "name": "energy_star",
50
+ "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark",
51
+ "dataset_name": "EnergyStarAI/text_generation",
52
+ "dataset_config": "",
53
+ "dataset_split": "train",
54
+ "num_samples": 1000,
55
+ "input_shapes": {
56
+ "batch_size": 1
57
+ },
58
+ "text_column_name": "text",
59
+ "truncation": true,
60
+ "max_length": -1,
61
+ "dataset_prefix1": "",
62
+ "dataset_prefix2": "",
63
+ "t5_task": "",
64
+ "image_column_name": "image",
65
+ "resize": false,
66
+ "question_column_name": "question",
67
+ "context_column_name": "context",
68
+ "sentence1_column_name": "sentence1",
69
+ "sentence2_column_name": "sentence2",
70
+ "audio_column_name": "audio",
71
+ "iterations": 10,
72
+ "warmup_runs": 10,
73
+ "energy": true,
74
+ "forward_kwargs": {},
75
+ "generate_kwargs": {
76
+ "max_new_tokens": 10,
77
+ "min_new_tokens": 10
78
+ },
79
+ "call_kwargs": {}
80
+ },
81
+ "environment": {
82
+ "cpu": " AMD EPYC 7R32",
83
+ "cpu_count": 48,
84
+ "cpu_ram_mb": 200472.73984,
85
+ "system": "Linux",
86
+ "machine": "x86_64",
87
+ "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
88
+ "processor": "x86_64",
89
+ "python_version": "3.9.20",
90
+ "gpu": [
91
+ "NVIDIA A10G"
92
+ ],
93
+ "gpu_count": 1,
94
+ "gpu_vram_mb": 24146608128,
95
+ "optimum_benchmark_version": "0.2.0",
96
+ "optimum_benchmark_commit": null,
97
+ "transformers_version": "4.44.0",
98
+ "transformers_commit": null,
99
+ "accelerate_version": "0.33.0",
100
+ "accelerate_commit": null,
101
+ "diffusers_version": "0.30.0",
102
+ "diffusers_commit": null,
103
+ "optimum_version": null,
104
+ "optimum_commit": null,
105
+ "timm_version": null,
106
+ "timm_commit": null,
107
+ "peft_version": null,
108
+ "peft_commit": null
109
+ }
110
+ }
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/generate_codecarbon.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "timestamp": "2024-10-25T13:52:02",
3
+ "project_name": "codecarbon",
4
+ "run_id": "8bd2e04f-cfc4-4707-8676-e5f40e6d32e7",
5
+ "duration": -1729480256.4145586,
6
+ "emissions": 0.005216170007237689,
7
+ "emissions_rate": 2.5454396948743776e-05,
8
+ "cpu_power": 42.5,
9
+ "gpu_power": 205.30973435218456,
10
+ "ram_power": 0.43064117431640625,
11
+ "cpu_energy": 0.0024192911417903404,
12
+ "gpu_energy": 0.011687011294046368,
13
+ "ram_energy": 2.4513656241329765e-05,
14
+ "energy_consumed": 0.014130816092078058,
15
+ "country_name": "United States",
16
+ "country_iso_code": "USA",
17
+ "region": "virginia",
18
+ "cloud_provider": "",
19
+ "cloud_region": "",
20
+ "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
21
+ "python_version": "3.9.20",
22
+ "codecarbon_version": "2.5.1",
23
+ "cpu_count": 48,
24
+ "cpu_model": "AMD EPYC 7R32",
25
+ "gpu_count": 1,
26
+ "gpu_model": "1 x NVIDIA A10G",
27
+ "longitude": -77.4903,
28
+ "latitude": 39.0469,
29
+ "ram_total_size": 186.7047882080078,
30
+ "tracking_mode": "process",
31
+ "on_cloud": "N",
32
+ "pue": 1.0
33
+ }
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/prefill_codecarbon.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "timestamp": "2024-10-25T13:17:51",
3
+ "project_name": "codecarbon",
4
+ "run_id": "8bd2e04f-cfc4-4707-8676-e5f40e6d32e7",
5
+ "duration": -1729480397.2874951,
6
+ "emissions": 0.0021972070187331605,
7
+ "emissions_rate": 3.430497464397154e-05,
8
+ "cpu_power": 42.5,
9
+ "gpu_power": 291.60627364641437,
10
+ "ram_power": 0.4305853843688965,
11
+ "cpu_energy": 0.000756214538868598,
12
+ "gpu_energy": 0.005188447206309732,
13
+ "ram_energy": 7.661176593828245e-06,
14
+ "energy_consumed": 0.0059523229217721615,
15
+ "country_name": "United States",
16
+ "country_iso_code": "USA",
17
+ "region": "virginia",
18
+ "cloud_provider": "",
19
+ "cloud_region": "",
20
+ "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
21
+ "python_version": "3.9.20",
22
+ "codecarbon_version": "2.5.1",
23
+ "cpu_count": 48,
24
+ "cpu_model": "AMD EPYC 7R32",
25
+ "gpu_count": 1,
26
+ "gpu_model": "1 x NVIDIA A10G",
27
+ "longitude": -77.4903,
28
+ "latitude": 39.0469,
29
+ "ram_total_size": 186.7047882080078,
30
+ "tracking_mode": "process",
31
+ "on_cloud": "N",
32
+ "pue": 1.0
33
+ }
text_generation/EleutherAI/pythia-1.4b/2024-10-25-13-06-57/preprocess_codecarbon.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "timestamp": "2024-10-25T13:07:11",
3
+ "project_name": "codecarbon",
4
+ "run_id": "8bd2e04f-cfc4-4707-8676-e5f40e6d32e7",
5
+ "duration": -1729480460.397016,
6
+ "emissions": 1.100585373385212e-05,
7
+ "emissions_rate": 1.1711243963235281e-05,
8
+ "cpu_power": 42.5,
9
+ "gpu_power": 71.30390443126397,
10
+ "ram_power": 0.32120847702026367,
11
+ "cpu_energy": 1.1119852669475626e-05,
12
+ "gpu_energy": 1.861168155592452e-05,
13
+ "ram_energy": 8.376988347301967e-08,
14
+ "energy_consumed": 2.9815304108873166e-05,
15
+ "country_name": "United States",
16
+ "country_iso_code": "USA",
17
+ "region": "virginia",
18
+ "cloud_provider": "",
19
+ "cloud_region": "",
20
+ "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
21
+ "python_version": "3.9.20",
22
+ "codecarbon_version": "2.5.1",
23
+ "cpu_count": 48,
24
+ "cpu_model": "AMD EPYC 7R32",
25
+ "gpu_count": 1,
26
+ "gpu_model": "1 x NVIDIA A10G",
27
+ "longitude": -77.4903,
28
+ "latitude": 39.0469,
29
+ "ram_total_size": 186.7047882080078,
30
+ "tracking_mode": "process",
31
+ "on_cloud": "N",
32
+ "pue": 1.0
33
+ }
text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48/.hydra/config.yaml ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ backend:
2
+ name: pytorch
3
+ version: 2.4.0
4
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
+ task: text-generation
6
+ model: meta-llama/Llama-3.1-8B-Instruct
7
+ processor: meta-llama/Llama-3.1-8B-Instruct
8
+ library: null
9
+ device: cuda
10
+ device_ids: '0'
11
+ seed: 42
12
+ inter_op_num_threads: null
13
+ intra_op_num_threads: null
14
+ hub_kwargs: {}
15
+ no_weights: true
16
+ device_map: null
17
+ torch_dtype: null
18
+ amp_autocast: false
19
+ amp_dtype: null
20
+ eval_mode: true
21
+ to_bettertransformer: false
22
+ low_cpu_mem_usage: null
23
+ attn_implementation: null
24
+ cache_implementation: null
25
+ torch_compile: false
26
+ torch_compile_config: {}
27
+ quantization_scheme: null
28
+ quantization_config: {}
29
+ deepspeed_inference: false
30
+ deepspeed_inference_config: {}
31
+ peft_type: null
32
+ peft_config: {}
33
+ launcher:
34
+ name: process
35
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
36
+ device_isolation: false
37
+ device_isolation_action: warn
38
+ start_method: spawn
39
+ benchmark:
40
+ name: energy_star
41
+ _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark
42
+ dataset_name: EnergyStarAI/text_generation
43
+ dataset_config: ''
44
+ dataset_split: train
45
+ num_samples: 1000
46
+ input_shapes:
47
+ batch_size: 1
48
+ text_column_name: text
49
+ truncation: true
50
+ max_length: -1
51
+ dataset_prefix1: ''
52
+ dataset_prefix2: ''
53
+ t5_task: ''
54
+ image_column_name: image
55
+ resize: false
56
+ question_column_name: question
57
+ context_column_name: context
58
+ sentence1_column_name: sentence1
59
+ sentence2_column_name: sentence2
60
+ audio_column_name: audio
61
+ iterations: 10
62
+ warmup_runs: 10
63
+ energy: true
64
+ forward_kwargs: {}
65
+ generate_kwargs:
66
+ max_new_tokens: 10
67
+ min_new_tokens: 10
68
+ call_kwargs: {}
69
+ experiment_name: text_generation
70
+ environment:
71
+ cpu: ' AMD EPYC 7R32'
72
+ cpu_count: 48
73
+ cpu_ram_mb: 200472.73984
74
+ system: Linux
75
+ machine: x86_64
76
+ platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
77
+ processor: x86_64
78
+ python_version: 3.9.20
79
+ gpu:
80
+ - NVIDIA A10G
81
+ gpu_count: 1
82
+ gpu_vram_mb: 24146608128
83
+ optimum_benchmark_version: 0.2.0
84
+ optimum_benchmark_commit: null
85
+ transformers_version: 4.44.0
86
+ transformers_commit: null
87
+ accelerate_version: 0.33.0
88
+ accelerate_commit: null
89
+ diffusers_version: 0.30.0
90
+ diffusers_commit: null
91
+ optimum_version: null
92
+ optimum_commit: null
93
+ timm_version: null
94
+ timm_commit: null
95
+ peft_version: null
96
+ peft_commit: null
text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48/.hydra/hydra.yaml ADDED
@@ -0,0 +1,175 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: /runs/text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48
4
+ sweep:
5
+ dir: sweeps/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ colorlog:
72
+ (): colorlog.ColoredFormatter
73
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
74
+ handlers:
75
+ console:
76
+ class: logging.StreamHandler
77
+ formatter: colorlog
78
+ stream: ext://sys.stdout
79
+ root:
80
+ level: INFO
81
+ handlers:
82
+ - console
83
+ disable_existing_loggers: false
84
+ job_logging:
85
+ version: 1
86
+ formatters:
87
+ simple:
88
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
89
+ colorlog:
90
+ (): colorlog.ColoredFormatter
91
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
92
+ - %(message)s'
93
+ log_colors:
94
+ DEBUG: purple
95
+ INFO: green
96
+ WARNING: yellow
97
+ ERROR: red
98
+ CRITICAL: red
99
+ handlers:
100
+ console:
101
+ class: logging.StreamHandler
102
+ formatter: colorlog
103
+ stream: ext://sys.stdout
104
+ file:
105
+ class: logging.FileHandler
106
+ formatter: simple
107
+ filename: ${hydra.job.name}.log
108
+ root:
109
+ level: INFO
110
+ handlers:
111
+ - console
112
+ - file
113
+ disable_existing_loggers: false
114
+ env: {}
115
+ mode: RUN
116
+ searchpath: []
117
+ callbacks: {}
118
+ output_subdir: .hydra
119
+ overrides:
120
+ hydra:
121
+ - hydra.run.dir=/runs/text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48
122
+ - hydra.mode=RUN
123
+ task:
124
+ - backend.model=meta-llama/Llama-3.1-8B-Instruct
125
+ - backend.processor=meta-llama/Llama-3.1-8B-Instruct
126
+ job:
127
+ name: cli
128
+ chdir: true
129
+ override_dirname: backend.model=meta-llama/Llama-3.1-8B-Instruct,backend.processor=meta-llama/Llama-3.1-8B-Instruct
130
+ id: ???
131
+ num: ???
132
+ config_name: text_generation
133
+ env_set:
134
+ OVERRIDE_BENCHMARKS: '1'
135
+ env_copy: []
136
+ config:
137
+ override_dirname:
138
+ kv_sep: '='
139
+ item_sep: ','
140
+ exclude_keys: []
141
+ runtime:
142
+ version: 1.3.2
143
+ version_base: '1.3'
144
+ cwd: /
145
+ config_sources:
146
+ - path: hydra.conf
147
+ schema: pkg
148
+ provider: hydra
149
+ - path: optimum_benchmark
150
+ schema: pkg
151
+ provider: main
152
+ - path: hydra_plugins.hydra_colorlog.conf
153
+ schema: pkg
154
+ provider: hydra-colorlog
155
+ - path: /optimum-benchmark/examples/energy_star
156
+ schema: file
157
+ provider: command-line
158
+ - path: ''
159
+ schema: structured
160
+ provider: schema
161
+ output_dir: /runs/text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48
162
+ choices:
163
+ benchmark: energy_star
164
+ launcher: process
165
+ backend: pytorch
166
+ hydra/env: default
167
+ hydra/callbacks: null
168
+ hydra/job_logging: colorlog
169
+ hydra/hydra_logging: colorlog
170
+ hydra/hydra_help: default
171
+ hydra/help: default
172
+ hydra/sweeper: basic
173
+ hydra/launcher: basic
174
+ hydra/output: default
175
+ verbose: false
text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48/.hydra/overrides.yaml ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ - backend.model=meta-llama/Llama-3.1-8B-Instruct
2
+ - backend.processor=meta-llama/Llama-3.1-8B-Instruct
text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48/cli.log ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2024-10-25 13:06:51,685][launcher][INFO] - ََAllocating process launcher
2
+ [2024-10-25 13:06:51,685][process][INFO] - + Setting multiprocessing start method to spawn.
3
+ [2024-10-25 13:06:51,696][process][INFO] - + Launched benchmark in isolated process 180.
4
+ [PROC-0][2024-10-25 13:06:54,217][datasets][INFO] - PyTorch version 2.4.0 available.
5
+ [PROC-0][2024-10-25 13:06:55,114][backend][INFO] - َAllocating pytorch backend
6
+ [PROC-0][2024-10-25 13:06:55,114][backend][INFO] - + Setting random seed to 42
7
+ [PROC-0][2024-10-25 13:06:56,109][pytorch][INFO] - + Using AutoModel class AutoModelForCausalLM
8
+ [PROC-0][2024-10-25 13:06:56,109][pytorch][INFO] - + Creating backend temporary directory
9
+ [PROC-0][2024-10-25 13:06:56,109][pytorch][INFO] - + Loading model with random weights
10
+ [PROC-0][2024-10-25 13:06:56,109][pytorch][INFO] - + Creating no weights model
11
+ [PROC-0][2024-10-25 13:06:56,109][pytorch][INFO] - + Creating no weights model directory
12
+ [PROC-0][2024-10-25 13:06:56,109][pytorch][INFO] - + Creating no weights model state dict
13
+ [PROC-0][2024-10-25 13:06:56,130][pytorch][INFO] - + Saving no weights model safetensors
14
+ [PROC-0][2024-10-25 13:06:56,131][pytorch][INFO] - + Saving no weights model pretrained config
15
+ [PROC-0][2024-10-25 13:06:56,131][pytorch][INFO] - + Loading no weights AutoModel
16
+ [PROC-0][2024-10-25 13:06:56,131][pytorch][INFO] - + Loading model directly on device: cuda
17
+ [2024-10-25 13:06:57,114][experiment][ERROR] - Error during experiment
text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48/error.log ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Error executing job with overrides: ['backend.model=meta-llama/Llama-3.1-8B-Instruct', 'backend.processor=meta-llama/Llama-3.1-8B-Instruct']
2
+ Traceback (most recent call last):
3
+ File "/optimum-benchmark/optimum_benchmark/cli.py", line 65, in benchmark_cli
4
+ benchmark_report: BenchmarkReport = launch(experiment_config=experiment_config)
5
+ File "/optimum-benchmark/optimum_benchmark/experiment.py", line 102, in launch
6
+ raise error
7
+ File "/optimum-benchmark/optimum_benchmark/experiment.py", line 90, in launch
8
+ report = launcher.launch(run, experiment_config.benchmark, experiment_config.backend)
9
+ File "/optimum-benchmark/optimum_benchmark/launchers/process/launcher.py", line 47, in launch
10
+ while not process_context.join():
11
+ File "/opt/conda/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 189, in join
12
+ raise ProcessRaisedException(msg, error_index, failed_process.pid)
13
+ torch.multiprocessing.spawn.ProcessRaisedException:
14
+
15
+ -- Process 0 terminated with the following error:
16
+ Traceback (most recent call last):
17
+ File "/opt/conda/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 76, in _wrap
18
+ fn(i, *args)
19
+ File "/optimum-benchmark/optimum_benchmark/launchers/process/launcher.py", line 63, in entrypoint
20
+ worker_output = worker(*worker_args)
21
+ File "/optimum-benchmark/optimum_benchmark/experiment.py", line 55, in run
22
+ backend: Backend = backend_factory(backend_config)
23
+ File "/optimum-benchmark/optimum_benchmark/backends/pytorch/backend.py", line 81, in __init__
24
+ self.load_model_with_no_weights()
25
+ File "/optimum-benchmark/optimum_benchmark/backends/pytorch/backend.py", line 246, in load_model_with_no_weights
26
+ self.load_model_from_pretrained()
27
+ File "/optimum-benchmark/optimum_benchmark/backends/pytorch/backend.py", line 204, in load_model_from_pretrained
28
+ self.pretrained_model = self.automodel_class.from_pretrained(
29
+ File "/opt/conda/lib/python3.9/site-packages/transformers/models/auto/auto_factory.py", line 564, in from_pretrained
30
+ return model_class.from_pretrained(
31
+ File "/opt/conda/lib/python3.9/site-packages/transformers/modeling_utils.py", line 3810, in from_pretrained
32
+ model = cls(config, *model_args, **model_kwargs)
33
+ File "/opt/conda/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 1116, in __init__
34
+ self.model = LlamaModel(config)
35
+ File "/opt/conda/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 902, in __init__
36
+ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)]
37
+ File "/opt/conda/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 902, in <listcomp>
38
+ [LlamaDecoderLayer(config, layer_idx) for layer_idx in range(config.num_hidden_layers)]
39
+ File "/opt/conda/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 691, in __init__
40
+ self.mlp = LlamaMLP(config)
41
+ File "/opt/conda/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 287, in __init__
42
+ self.up_proj = nn.Linear(self.hidden_size, self.intermediate_size, bias=config.mlp_bias)
43
+ File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/linear.py", line 99, in __init__
44
+ self.weight = Parameter(torch.empty((out_features, in_features), **factory_kwargs))
45
+ File "/opt/conda/lib/python3.9/site-packages/torch/utils/_device.py", line 79, in __torch_function__
46
+ return func(*args, **kwargs)
47
+ torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU 0 has a total capacity of 22.19 GiB of which 69.50 MiB is free. Process 378708 has 22.12 GiB memory in use. Of the allocated memory 21.83 GiB is allocated by PyTorch, and 1.24 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
48
+
49
+
50
+ Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
text_generation/meta-llama/Llama-3.1-8B-Instruct/2024-10-25-13-06-48/experiment_config.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_name": "text_generation",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "model": "meta-llama/Llama-3.1-8B-Instruct",
9
+ "processor": "meta-llama/Llama-3.1-8B-Instruct",
10
+ "library": "transformers",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "hub_kwargs": {
17
+ "revision": "main",
18
+ "force_download": false,
19
+ "local_files_only": false,
20
+ "trust_remote_code": true
21
+ },
22
+ "no_weights": true,
23
+ "device_map": null,
24
+ "torch_dtype": null,
25
+ "amp_autocast": false,
26
+ "amp_dtype": null,
27
+ "eval_mode": true,
28
+ "to_bettertransformer": false,
29
+ "low_cpu_mem_usage": null,
30
+ "attn_implementation": null,
31
+ "cache_implementation": null,
32
+ "torch_compile": false,
33
+ "torch_compile_config": {},
34
+ "quantization_scheme": null,
35
+ "quantization_config": {},
36
+ "deepspeed_inference": false,
37
+ "deepspeed_inference_config": {},
38
+ "peft_type": null,
39
+ "peft_config": {}
40
+ },
41
+ "launcher": {
42
+ "name": "process",
43
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
44
+ "device_isolation": false,
45
+ "device_isolation_action": "warn",
46
+ "start_method": "spawn"
47
+ },
48
+ "benchmark": {
49
+ "name": "energy_star",
50
+ "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark",
51
+ "dataset_name": "EnergyStarAI/text_generation",
52
+ "dataset_config": "",
53
+ "dataset_split": "train",
54
+ "num_samples": 1000,
55
+ "input_shapes": {
56
+ "batch_size": 1
57
+ },
58
+ "text_column_name": "text",
59
+ "truncation": true,
60
+ "max_length": -1,
61
+ "dataset_prefix1": "",
62
+ "dataset_prefix2": "",
63
+ "t5_task": "",
64
+ "image_column_name": "image",
65
+ "resize": false,
66
+ "question_column_name": "question",
67
+ "context_column_name": "context",
68
+ "sentence1_column_name": "sentence1",
69
+ "sentence2_column_name": "sentence2",
70
+ "audio_column_name": "audio",
71
+ "iterations": 10,
72
+ "warmup_runs": 10,
73
+ "energy": true,
74
+ "forward_kwargs": {},
75
+ "generate_kwargs": {
76
+ "max_new_tokens": 10,
77
+ "min_new_tokens": 10
78
+ },
79
+ "call_kwargs": {}
80
+ },
81
+ "environment": {
82
+ "cpu": " AMD EPYC 7R32",
83
+ "cpu_count": 48,
84
+ "cpu_ram_mb": 200472.73984,
85
+ "system": "Linux",
86
+ "machine": "x86_64",
87
+ "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
88
+ "processor": "x86_64",
89
+ "python_version": "3.9.20",
90
+ "gpu": [
91
+ "NVIDIA A10G"
92
+ ],
93
+ "gpu_count": 1,
94
+ "gpu_vram_mb": 24146608128,
95
+ "optimum_benchmark_version": "0.2.0",
96
+ "optimum_benchmark_commit": null,
97
+ "transformers_version": "4.44.0",
98
+ "transformers_commit": null,
99
+ "accelerate_version": "0.33.0",
100
+ "accelerate_commit": null,
101
+ "diffusers_version": "0.30.0",
102
+ "diffusers_commit": null,
103
+ "optimum_version": null,
104
+ "optimum_commit": null,
105
+ "timm_version": null,
106
+ "timm_commit": null,
107
+ "peft_version": null,
108
+ "peft_commit": null
109
+ }
110
+ }