The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code:   DatasetGenerationError
Exception:    ArrowNotImplementedError
Message:      Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1869, in _prepare_split_single
                  writer.write_table(table)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 578, in write_table
                  self._build_writer(inferred_schema=pa_table.schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 399, in _build_writer
                  self.pa_writer = self._WRITER_CLASS(self.stream, schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__
                  self.writer = _parquet.ParquetWriter(
                File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1885, in _prepare_split_single
                  num_examples, num_bytes = writer.finalize()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 597, in finalize
                  self._build_writer(self.schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 399, in _build_writer
                  self.pa_writer = self._WRITER_CLASS(self.stream, schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__
                  self.writer = _parquet.ParquetWriter(
                File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1392, in compute_config_parquet_and_info_response
                  parquet_operations = convert_to_parquet(builder)
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1041, in convert_to_parquet
                  builder.download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 924, in download_and_prepare
                  self._download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 999, in _download_and_prepare
                  self._prepare_split(split_generator, **prepare_split_kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1740, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1896, in _prepare_split_single
                  raise DatasetGenerationError("An error occurred while generating the dataset") from e
              datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

config
dict
report
dict
name
string
backend
dict
scenario
dict
launcher
dict
environment
dict
print_report
bool
log_report
bool
overall
dict
warmup
dict
train
dict
{ "name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "library": "transformers", "model_type": "bert", "model": "google-bert/bert-base-uncased", "processor": "google-bert/bert-base-uncased", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1299.668992, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "values": [ 0.3378821105957031, 0.043873279571533204, 0.04286873626708984, 0.04303155136108398, 0.04216320037841797 ], "count": 5, "total": 0.5098188781738281, "mean": 0.10196377563476562, "p50": 0.04303155136108398, "p90": 0.2202785781860352, "p95": 0.2790803443908691, "p99": 0.3261217573547363, "stdev": 0.1179604222104298, "stdev_": 115.68855848665724 }, "throughput": { "unit": "samples/s", "value": 98.07404578484827 }, "energy": { "unit": "kWh", "cpu": 0.00001023778893402686, "ram": 0.000005588101379470862, "gpu": 0.000017930014344001016, "total": 0.00003375590465749874 }, "efficiency": { "unit": "samples/kWh", "value": 296244.4674928462 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 1299.668992, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "values": [ 0.3378821105957031, 0.043873279571533204 ], "count": 2, "total": 0.3817553901672363, "mean": 0.19087769508361815, "p50": 0.19087769508361815, "p90": 0.3084812274932861, "p95": 0.3231816690444946, "p99": 0.3349420222854614, "stdev": 0.14700441551208496, "stdev_": 77.01497833347499 }, "throughput": { "unit": "samples/s", "value": 20.95582722877973 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1299.668992, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "values": [ 0.04286873626708984, 0.04303155136108398, 0.04216320037841797 ], "count": 3, "total": 0.12806348800659179, "mean": 0.04268782933553059, "p50": 0.04286873626708984, "p90": 0.042998988342285153, "p95": 0.04301526985168457, "p99": 0.0430282950592041, "stdev": 0.0003768765011591056, "stdev_": 0.8828663978128725 }, "throughput": { "unit": "samples/s", "value": 140.55528457161412 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_fill-mask_google-bert/bert-base-uncased
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "library": "transformers", "model_type": "bert", "model": "google-bert/bert-base-uncased", "processor": "google-bert/bert-base-uncased", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1299.668992, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "values": [ 0.3378821105957031, 0.043873279571533204, 0.04286873626708984, 0.04303155136108398, 0.04216320037841797 ], "count": 5, "total": 0.5098188781738281, "mean": 0.10196377563476562, "p50": 0.04303155136108398, "p90": 0.2202785781860352, "p95": 0.2790803443908691, "p99": 0.3261217573547363, "stdev": 0.1179604222104298, "stdev_": 115.68855848665724 }, "throughput": { "unit": "samples/s", "value": 98.07404578484827 }, "energy": { "unit": "kWh", "cpu": 0.00001023778893402686, "ram": 0.000005588101379470862, "gpu": 0.000017930014344001016, "total": 0.00003375590465749874 }, "efficiency": { "unit": "samples/kWh", "value": 296244.4674928462 } }
{ "memory": { "unit": "MB", "max_ram": 1299.668992, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "values": [ 0.3378821105957031, 0.043873279571533204 ], "count": 2, "total": 0.3817553901672363, "mean": 0.19087769508361815, "p50": 0.19087769508361815, "p90": 0.3084812274932861, "p95": 0.3231816690444946, "p99": 0.3349420222854614, "stdev": 0.14700441551208496, "stdev_": 77.01497833347499 }, "throughput": { "unit": "samples/s", "value": 20.95582722877973 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1299.668992, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "values": [ 0.04286873626708984, 0.04303155136108398, 0.04216320037841797 ], "count": 3, "total": 0.12806348800659179, "mean": 0.04268782933553059, "p50": 0.04286873626708984, "p90": 0.042998988342285153, "p95": 0.04301526985168457, "p99": 0.0430282950592041, "stdev": 0.0003768765011591056, "stdev_": 0.8828663978128725 }, "throughput": { "unit": "samples/s", "value": 140.55528457161412 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "model": "google-bert/bert-base-uncased", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1063.8336, "max_global_vram": 3169.32096, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 5, "total": 0.7448790740966797, "mean": 0.14897581481933594, "stdev": 0.2054173633207176, "p50": 0.04632883071899414, "p90": 0.35471870422363283, "p95": 0.4572641067504882, "p99": 0.5393004287719726, "values": [ 0.5598095092773437, 0.04708249664306641, 0.04632883071899414, 0.04576665496826172, 0.04589158248901367 ] }, "throughput": { "unit": "samples/s", "value": 67.1249894630687 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1063.8336, "max_global_vram": 3169.32096, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 2, "total": 0.6068920059204101, "mean": 0.3034460029602051, "stdev": 0.25636350631713867, "p50": 0.3034460029602051, "p90": 0.508536808013916, "p95": 0.5341731586456299, "p99": 0.554682239151001, "values": [ 0.5598095092773437, 0.04708249664306641 ] }, "throughput": { "unit": "samples/s", "value": 13.181916917602548 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1063.8336, "max_global_vram": 3169.32096, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 3, "total": 0.13798706817626955, "mean": 0.045995689392089846, "stdev": 0.00024102431292157434, "p50": 0.04589158248901367, "p90": 0.04624138107299805, "p95": 0.0462851058959961, "p99": 0.04632008575439454, "values": [ 0.04632883071899414, 0.04576665496826172, 0.04589158248901367 ] }, "throughput": { "unit": "samples/s", "value": 130.4470066499722 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "library": "transformers", "model_type": "vit", "model": "google/vit-base-patch16-224", "processor": "google/vit-base-patch16-224", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1631.678464, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "values": [ 0.3338004455566406, 0.04012441635131836, 0.039790592193603515, 0.03977519989013672, 0.039820289611816405 ], "count": 5, "total": 0.49331094360351563, "mean": 0.09866218872070312, "p50": 0.039820289611816405, "p90": 0.21633003387451175, "p95": 0.2750652397155761, "p99": 0.3220534043884277, "stdev": 0.11756919838438508, "stdev_": 119.1633795163461 }, "throughput": { "unit": "samples/s", "value": 101.3559513493908 }, "energy": { "unit": "kWh", "cpu": 0.000011456950427083257, "ram": 0.000006241313718108779, "gpu": 0.000023570574411999404, "total": 0.00004126883855719144 }, "efficiency": { "unit": "samples/kWh", "value": 242313.5796792957 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 1631.678464, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "values": [ 0.3338004455566406, 0.04012441635131836 ], "count": 2, "total": 0.373924861907959, "mean": 0.1869624309539795, "p50": 0.1869624309539795, "p90": 0.3044328426361084, "p95": 0.3191166440963745, "p99": 0.3308636852645874, "stdev": 0.14683801460266113, "stdev_": 78.53878121578612 }, "throughput": { "unit": "samples/s", "value": 21.39467260662969 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1631.678464, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "values": [ 0.039790592193603515, 0.03977519989013672, 0.039820289611816405 ], "count": 3, "total": 0.11938608169555665, "mean": 0.039795360565185546, "p50": 0.039790592193603515, "p90": 0.039814350128173824, "p95": 0.039817319869995114, "p99": 0.03981969566345215, "stdev": 0.000018714054895472507, "stdev_": 0.04702572015855601 }, "throughput": { "unit": "samples/s", "value": 150.77134406589653 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_image-classification_google/vit-base-patch16-224
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "library": "transformers", "model_type": "vit", "model": "google/vit-base-patch16-224", "processor": "google/vit-base-patch16-224", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1631.678464, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "values": [ 0.3338004455566406, 0.04012441635131836, 0.039790592193603515, 0.03977519989013672, 0.039820289611816405 ], "count": 5, "total": 0.49331094360351563, "mean": 0.09866218872070312, "p50": 0.039820289611816405, "p90": 0.21633003387451175, "p95": 0.2750652397155761, "p99": 0.3220534043884277, "stdev": 0.11756919838438508, "stdev_": 119.1633795163461 }, "throughput": { "unit": "samples/s", "value": 101.3559513493908 }, "energy": { "unit": "kWh", "cpu": 0.000011456950427083257, "ram": 0.000006241313718108779, "gpu": 0.000023570574411999404, "total": 0.00004126883855719144 }, "efficiency": { "unit": "samples/kWh", "value": 242313.5796792957 } }
{ "memory": { "unit": "MB", "max_ram": 1631.678464, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "values": [ 0.3338004455566406, 0.04012441635131836 ], "count": 2, "total": 0.373924861907959, "mean": 0.1869624309539795, "p50": 0.1869624309539795, "p90": 0.3044328426361084, "p95": 0.3191166440963745, "p99": 0.3308636852645874, "stdev": 0.14683801460266113, "stdev_": 78.53878121578612 }, "throughput": { "unit": "samples/s", "value": 21.39467260662969 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1631.678464, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "values": [ 0.039790592193603515, 0.03977519989013672, 0.039820289611816405 ], "count": 3, "total": 0.11938608169555665, "mean": 0.039795360565185546, "p50": 0.039790592193603515, "p90": 0.039814350128173824, "p95": 0.039817319869995114, "p99": 0.03981969566345215, "stdev": 0.000018714054895472507, "stdev_": 0.04702572015855601 }, "throughput": { "unit": "samples/s", "value": 150.77134406589653 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "model": "google/vit-base-patch16-224", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1446.354944, "max_global_vram": 2628.255744, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1756.126208 }, "latency": { "unit": "s", "count": 5, "total": 0.48406525039672854, "mean": 0.09681305007934571, "stdev": 0.1110534407118009, "p50": 0.04146995162963867, "p90": 0.20794796142578126, "p95": 0.26343380432128904, "p99": 0.3078224786376953, "values": [ 0.3189196472167969, 0.04103168106079102, 0.041490432739257815, 0.04146995162963867, 0.04115353775024414 ] }, "throughput": { "unit": "samples/s", "value": 103.29185984538483 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1446.354944, "max_global_vram": 2628.255744, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1756.126208 }, "latency": { "unit": "s", "count": 2, "total": 0.3599513282775879, "mean": 0.17997566413879396, "stdev": 0.13894398307800293, "p50": 0.17997566413879396, "p90": 0.2911308506011963, "p95": 0.30502524890899657, "p99": 0.31614076755523685, "values": [ 0.3189196472167969, 0.04103168106079102 ] }, "throughput": { "unit": "samples/s", "value": 22.225227055782792 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1446.354944, "max_global_vram": 2628.255744, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1756.126208 }, "latency": { "unit": "s", "count": 3, "total": 0.12411392211914063, "mean": 0.041371307373046874, "stdev": 0.00015421321911462263, "p50": 0.04146995162963867, "p90": 0.04148633651733399, "p95": 0.0414883846282959, "p99": 0.041490023117065435, "values": [ 0.041490432739257815, 0.04146995162963867, 0.04115353775024414 ] }, "throughput": { "unit": "samples/s", "value": 145.02804917180256 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1337.192448, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "values": [ 0.35576217651367187, 0.04703129577636719, 0.04605235290527344, 0.046074878692626955, 0.04605542373657227 ], "count": 5, "total": 0.5409761276245116, "mean": 0.10819522552490232, "p50": 0.046074878692626955, "p90": 0.23226982421875003, "p95": 0.2940160003662109, "p99": 0.34341294128417965, "stdev": 0.12378404630519566, "stdev_": 114.40804869592456 }, "throughput": { "unit": "samples/s", "value": 92.42552017878451 }, "energy": { "unit": "kWh", "cpu": 0.00001055118112430383, "ram": 0.0000057624615054751955, "gpu": 0.000017052791419999887, "total": 0.00003336643404977891 }, "efficiency": { "unit": "samples/kWh", "value": 299702.389086024 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 1337.192448, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "values": [ 0.35576217651367187, 0.04703129577636719 ], "count": 2, "total": 0.40279347229003903, "mean": 0.20139673614501952, "p50": 0.20139673614501952, "p90": 0.3248890884399414, "p95": 0.34032563247680664, "p99": 0.3526748677062988, "stdev": 0.15436544036865232, "stdev_": 76.64743894235635 }, "throughput": { "unit": "samples/s", "value": 19.86129505653818 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1337.192448, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "values": [ 0.04605235290527344, 0.046074878692626955, 0.04605542373657227 ], "count": 3, "total": 0.13818265533447266, "mean": 0.04606088511149089, "p50": 0.04605542373657227, "p90": 0.046070987701416016, "p95": 0.046072933197021486, "p99": 0.04607448959350586, "stdev": 0.000009974057548636764, "stdev_": 0.021654072700718726 }, "throughput": { "unit": "samples/s", "value": 130.2623687208123 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_multiple-choice_FacebookAI/roberta-base
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1337.192448, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "values": [ 0.35576217651367187, 0.04703129577636719, 0.04605235290527344, 0.046074878692626955, 0.04605542373657227 ], "count": 5, "total": 0.5409761276245116, "mean": 0.10819522552490232, "p50": 0.046074878692626955, "p90": 0.23226982421875003, "p95": 0.2940160003662109, "p99": 0.34341294128417965, "stdev": 0.12378404630519566, "stdev_": 114.40804869592456 }, "throughput": { "unit": "samples/s", "value": 92.42552017878451 }, "energy": { "unit": "kWh", "cpu": 0.00001055118112430383, "ram": 0.0000057624615054751955, "gpu": 0.000017052791419999887, "total": 0.00003336643404977891 }, "efficiency": { "unit": "samples/kWh", "value": 299702.389086024 } }
{ "memory": { "unit": "MB", "max_ram": 1337.192448, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "values": [ 0.35576217651367187, 0.04703129577636719 ], "count": 2, "total": 0.40279347229003903, "mean": 0.20139673614501952, "p50": 0.20139673614501952, "p90": 0.3248890884399414, "p95": 0.34032563247680664, "p99": 0.3526748677062988, "stdev": 0.15436544036865232, "stdev_": 76.64743894235635 }, "throughput": { "unit": "samples/s", "value": 19.86129505653818 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1337.192448, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "values": [ 0.04605235290527344, 0.046074878692626955, 0.04605542373657227 ], "count": 3, "total": 0.13818265533447266, "mean": 0.04606088511149089, "p50": 0.04605542373657227, "p90": 0.046070987701416016, "p95": 0.046072933197021486, "p99": 0.04607448959350586, "stdev": 0.000009974057548636764, "stdev_": 0.021654072700718726 }, "throughput": { "unit": "samples/s", "value": 130.2623687208123 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "model": "FacebookAI/roberta-base", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1093.496832, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 5, "total": 0.8026234703063965, "mean": 0.16052469406127928, "stdev": 0.22240891148008993, "p50": 0.04907724761962891, "p90": 0.38326721343994147, "p95": 0.49430444412231433, "p99": 0.5831342286682129, "values": [ 0.6053416748046875, 0.05015552139282226, 0.04897484970092773, 0.04907417678833008, 0.04907724761962891 ] }, "throughput": { "unit": "samples/s", "value": 62.295711313940046 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1093.496832, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 2, "total": 0.6554971961975098, "mean": 0.3277485980987549, "stdev": 0.27759307670593264, "p50": 0.3277485980987549, "p90": 0.549823059463501, "p95": 0.5775823671340942, "p99": 0.5997898132705688, "values": [ 0.6053416748046875, 0.05015552139282226 ] }, "throughput": { "unit": "samples/s", "value": 12.204476306546239 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1093.496832, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 3, "total": 0.14712627410888673, "mean": 0.049042091369628914, "stdev": 0.000047563564546161045, "p50": 0.04907417678833008, "p90": 0.04907663345336914, "p95": 0.049076940536499025, "p99": 0.04907718620300293, "values": [ 0.04897484970092773, 0.04907417678833008, 0.04907724761962891 ] }, "throughput": { "unit": "samples/s", "value": 122.34388527149387 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1323.085824, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "values": [ 0.3465441284179688, 0.04596223831176758, 0.045655040740966796, 0.04518291091918945, 0.045499393463134766 ], "count": 5, "total": 0.5288437118530275, "mean": 0.10576874237060549, "p50": 0.045655040740966796, "p90": 0.22631137237548832, "p95": 0.2864277503967285, "p99": 0.33452085281372074, "stdev": 0.12038795535102645, "stdev_": 113.82186518697213 }, "throughput": { "unit": "samples/s", "value": 94.54589111933254 }, "energy": { "unit": "kWh", "cpu": 0.000010490307170139897, "ram": 0.000005714536570311505, "gpu": 0.000018112236711999644, "total": 0.00003431708045245105 }, "efficiency": { "unit": "samples/kWh", "value": 291400.08031440113 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 1323.085824, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "values": [ 0.3465441284179688, 0.04596223831176758 ], "count": 2, "total": 0.3925063667297364, "mean": 0.1962531833648682, "p50": 0.1962531833648682, "p90": 0.31648593940734865, "p95": 0.3315150339126587, "p99": 0.3435383095169068, "stdev": 0.1502909450531006, "stdev_": 76.58013107164945 }, "throughput": { "unit": "samples/s", "value": 20.38183499201293 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1323.085824, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "values": [ 0.045655040740966796, 0.04518291091918945, 0.045499393463134766 ], "count": 3, "total": 0.13633734512329101, "mean": 0.04544578170776367, "p50": 0.045499393463134766, "p90": 0.04562391128540039, "p95": 0.04563947601318359, "p99": 0.045651927795410156, "stdev": 0.00019643880707632385, "stdev_": 0.4322487141700227 }, "throughput": { "unit": "samples/s", "value": 132.02545482840705 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_text-classification_FacebookAI/roberta-base
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1323.085824, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "values": [ 0.3465441284179688, 0.04596223831176758, 0.045655040740966796, 0.04518291091918945, 0.045499393463134766 ], "count": 5, "total": 0.5288437118530275, "mean": 0.10576874237060549, "p50": 0.045655040740966796, "p90": 0.22631137237548832, "p95": 0.2864277503967285, "p99": 0.33452085281372074, "stdev": 0.12038795535102645, "stdev_": 113.82186518697213 }, "throughput": { "unit": "samples/s", "value": 94.54589111933254 }, "energy": { "unit": "kWh", "cpu": 0.000010490307170139897, "ram": 0.000005714536570311505, "gpu": 0.000018112236711999644, "total": 0.00003431708045245105 }, "efficiency": { "unit": "samples/kWh", "value": 291400.08031440113 } }
{ "memory": { "unit": "MB", "max_ram": 1323.085824, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "values": [ 0.3465441284179688, 0.04596223831176758 ], "count": 2, "total": 0.3925063667297364, "mean": 0.1962531833648682, "p50": 0.1962531833648682, "p90": 0.31648593940734865, "p95": 0.3315150339126587, "p99": 0.3435383095169068, "stdev": 0.1502909450531006, "stdev_": 76.58013107164945 }, "throughput": { "unit": "samples/s", "value": 20.38183499201293 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1323.085824, "max_global_vram": 3384.27904, "max_process_vram": 0, "max_reserved": 2728.394752, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "values": [ 0.045655040740966796, 0.04518291091918945, 0.045499393463134766 ], "count": 3, "total": 0.13633734512329101, "mean": 0.04544578170776367, "p50": 0.045499393463134766, "p90": 0.04562391128540039, "p95": 0.04563947601318359, "p99": 0.045651927795410156, "stdev": 0.00019643880707632385, "stdev_": 0.4322487141700227 }, "throughput": { "unit": "samples/s", "value": 132.02545482840705 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "model": "FacebookAI/roberta-base", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1080.1152, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 5, "total": 0.778464241027832, "mean": 0.1556928482055664, "stdev": 0.21129125718980005, "p50": 0.05020159912109375, "p90": 0.36730080566406254, "p95": 0.4727875488281249, "p99": 0.557176943359375, "values": [ 0.5782742919921875, 0.050840576171875, 0.04969574356079102, 0.049452030181884765, 0.05020159912109375 ] }, "throughput": { "unit": "samples/s", "value": 64.22902602948513 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1080.1152, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 2, "total": 0.6291148681640625, "mean": 0.31455743408203124, "stdev": 0.26371685791015625, "p50": 0.31455743408203124, "p90": 0.5255309204101563, "p95": 0.5519026062011718, "p99": 0.5729999548339844, "values": [ 0.5782742919921875, 0.050840576171875 ] }, "throughput": { "unit": "samples/s", "value": 12.716278703357135 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1080.1152, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 3, "total": 0.14934937286376954, "mean": 0.049783124287923176, "stdev": 0.0003121857804388603, "p50": 0.04969574356079102, "p90": 0.05010042800903321, "p95": 0.050151013565063476, "p99": 0.0501914820098877, "values": [ 0.04969574356079102, 0.049452030181884765, 0.05020159912109375 ] }, "throughput": { "unit": "samples/s", "value": 120.52276922795568 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_text-generation_openai-community/gpt2", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "library": "transformers", "model_type": "gpt2", "model": "openai-community/gpt2", "processor": "openai-community/gpt2", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1350.053888, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "values": [ 0.35609701538085936, 0.044365825653076174, 0.04398284912109375, 0.043514881134033206, 0.04412108612060547 ], "count": 5, "total": 0.532081657409668, "mean": 0.1064163314819336, "p50": 0.04412108612060547, "p90": 0.2314045394897461, "p95": 0.29375077743530265, "p99": 0.34362776779174803, "stdev": 0.12484064959463413, "stdev_": 117.3134309895173 }, "throughput": { "unit": "samples/s", "value": 93.97053873913808 }, "energy": { "unit": "kWh", "cpu": 0.00001046066729236246, "ram": 0.000005713810620613953, "gpu": 0.000015273067774000063, "total": 0.00003144754568697648 }, "efficiency": { "unit": "samples/kWh", "value": 317989.8393196817 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 1350.053888, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "values": [ 0.35609701538085936, 0.044365825653076174 ], "count": 2, "total": 0.40046284103393553, "mean": 0.20023142051696777, "p50": 0.20023142051696777, "p90": 0.32492389640808106, "p95": 0.3405104558944702, "p99": 0.35297970348358154, "stdev": 0.1558655948638916, "stdev_": 77.84272541315933 }, "throughput": { "unit": "samples/s", "value": 19.976884695082294 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1350.053888, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "values": [ 0.04398284912109375, 0.043514881134033206, 0.04412108612060547 ], "count": 3, "total": 0.13161881637573242, "mean": 0.04387293879191081, "p50": 0.04398284912109375, "p90": 0.04409343872070312, "p95": 0.0441072624206543, "p99": 0.044118321380615236, "stdev": 0.0002593984474151631, "stdev_": 0.5912493089316151 }, "throughput": { "unit": "samples/s", "value": 136.75856154651456 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_text-generation_openai-community/gpt2
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "library": "transformers", "model_type": "gpt2", "model": "openai-community/gpt2", "processor": "openai-community/gpt2", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1350.053888, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "values": [ 0.35609701538085936, 0.044365825653076174, 0.04398284912109375, 0.043514881134033206, 0.04412108612060547 ], "count": 5, "total": 0.532081657409668, "mean": 0.1064163314819336, "p50": 0.04412108612060547, "p90": 0.2314045394897461, "p95": 0.29375077743530265, "p99": 0.34362776779174803, "stdev": 0.12484064959463413, "stdev_": 117.3134309895173 }, "throughput": { "unit": "samples/s", "value": 93.97053873913808 }, "energy": { "unit": "kWh", "cpu": 0.00001046066729236246, "ram": 0.000005713810620613953, "gpu": 0.000015273067774000063, "total": 0.00003144754568697648 }, "efficiency": { "unit": "samples/kWh", "value": 317989.8393196817 } }
{ "memory": { "unit": "MB", "max_ram": 1350.053888, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "values": [ 0.35609701538085936, 0.044365825653076174 ], "count": 2, "total": 0.40046284103393553, "mean": 0.20023142051696777, "p50": 0.20023142051696777, "p90": 0.32492389640808106, "p95": 0.3405104558944702, "p99": 0.35297970348358154, "stdev": 0.1558655948638916, "stdev_": 77.84272541315933 }, "throughput": { "unit": "samples/s", "value": 19.976884695082294 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1350.053888, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "values": [ 0.04398284912109375, 0.043514881134033206, 0.04412108612060547 ], "count": 3, "total": 0.13161881637573242, "mean": 0.04387293879191081, "p50": 0.04398284912109375, "p90": 0.04409343872070312, "p95": 0.0441072624206543, "p99": 0.044118321380615236, "stdev": 0.0002593984474151631, "stdev_": 0.5912493089316151 }, "throughput": { "unit": "samples/s", "value": 136.75856154651456 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_text-generation_openai-community/gpt2", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "model": "openai-community/gpt2", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1107.173376, "max_global_vram": 3563.585536, "max_process_vram": 0, "max_reserved": 2915.04128, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 5, "total": 0.8139294586181639, "mean": 0.16278589172363278, "stdev": 0.2273662362653359, "p50": 0.04927385711669922, "p90": 0.3902293930053711, "p95": 0.5038737297058105, "p99": 0.594789199066162, "values": [ 0.61751806640625, 0.049296382904052735, 0.04860006332397461, 0.0492410888671875, 0.04927385711669922 ] }, "throughput": { "unit": "samples/s", "value": 61.4303849929289 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1107.173376, "max_global_vram": 3563.585536, "max_process_vram": 0, "max_reserved": 2915.04128, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 2, "total": 0.6668144493103026, "mean": 0.3334072246551513, "stdev": 0.2841108417510986, "p50": 0.3334072246551513, "p90": 0.5606958980560303, "p95": 0.5891069822311401, "p99": 0.6118358495712279, "values": [ 0.61751806640625, 0.049296382904052735 ] }, "throughput": { "unit": "samples/s", "value": 11.99734050195603 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1107.173376, "max_global_vram": 3563.585536, "max_process_vram": 0, "max_reserved": 2915.04128, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 3, "total": 0.1471150093078613, "mean": 0.04903833643595377, "stdev": 0.00031019448743967263, "p50": 0.0492410888671875, "p90": 0.04926730346679687, "p95": 0.049270580291748044, "p99": 0.04927320175170898, "values": [ 0.04860006332397461, 0.0492410888671875, 0.04927385711669922 ] }, "throughput": { "unit": "samples/s", "value": 122.35325331307405 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "library": "transformers", "model_type": "deberta-v2", "model": "microsoft/deberta-v3-base", "processor": "microsoft/deberta-v3-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }, "print_report": true, "log_report": true }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1386.917888, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "values": [ 0.4221747131347656, 0.09074073791503906, 0.08792063903808593, 0.08825856018066407, 0.08857084655761718 ], "count": 5, "total": 0.7776654968261717, "mean": 0.15553309936523435, "p50": 0.08857084655761718, "p90": 0.289601123046875, "p95": 0.35588791809082027, "p99": 0.4089173541259766, "stdev": 0.1333244553793267, "stdev_": 85.72095323982731 }, "throughput": { "unit": "samples/s", "value": 64.2949959899999 }, "energy": { "unit": "kWh", "cpu": 0.000013409542672222587, "ram": 0.000007298270686959867, "gpu": 0.00002345001876000001, "total": 0.00004415783211918246 }, "efficiency": { "unit": "samples/kWh", "value": 226460.39264359474 } }, "warmup": { "memory": { "unit": "MB", "max_ram": 1386.917888, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "values": [ 0.4221747131347656, 0.09074073791503906 ], "count": 2, "total": 0.5129154510498046, "mean": 0.2564577255249023, "p50": 0.2564577255249023, "p90": 0.38903131561279297, "p95": 0.40560301437377927, "p99": 0.4188603733825684, "stdev": 0.16571698760986328, "stdev_": 64.61766252924676 }, "throughput": { "unit": "samples/s", "value": 15.597112513623989 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1386.917888, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "values": [ 0.08792063903808593, 0.08825856018066407, 0.08857084655761718 ], "count": 3, "total": 0.2647500457763672, "mean": 0.08825001525878906, "p50": 0.08825856018066407, "p90": 0.08850838928222655, "p95": 0.08853961791992188, "p99": 0.08856460083007812, "stdev": 0.00026551486635231693, "stdev_": 0.3008666520608602 }, "throughput": { "unit": "samples/s", "value": 67.98865679972155 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_token-classification_microsoft/deberta-v3-base
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "library": "transformers", "model_type": "deberta-v2", "model": "microsoft/deberta-v3-base", "processor": "microsoft/deberta-v3-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": true }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": null, "transformers_version": "4.45.2", "transformers_commit": null, "accelerate_version": "1.0.1", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.13.2", "peft_commit": null }
true
true
null
null
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1386.917888, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "values": [ 0.4221747131347656, 0.09074073791503906, 0.08792063903808593, 0.08825856018066407, 0.08857084655761718 ], "count": 5, "total": 0.7776654968261717, "mean": 0.15553309936523435, "p50": 0.08857084655761718, "p90": 0.289601123046875, "p95": 0.35588791809082027, "p99": 0.4089173541259766, "stdev": 0.1333244553793267, "stdev_": 85.72095323982731 }, "throughput": { "unit": "samples/s", "value": 64.2949959899999 }, "energy": { "unit": "kWh", "cpu": 0.000013409542672222587, "ram": 0.000007298270686959867, "gpu": 0.00002345001876000001, "total": 0.00004415783211918246 }, "efficiency": { "unit": "samples/kWh", "value": 226460.39264359474 } }
{ "memory": { "unit": "MB", "max_ram": 1386.917888, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "values": [ 0.4221747131347656, 0.09074073791503906 ], "count": 2, "total": 0.5129154510498046, "mean": 0.2564577255249023, "p50": 0.2564577255249023, "p90": 0.38903131561279297, "p95": 0.40560301437377927, "p99": 0.4188603733825684, "stdev": 0.16571698760986328, "stdev_": 64.61766252924676 }, "throughput": { "unit": "samples/s", "value": 15.597112513623989 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1386.917888, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "values": [ 0.08792063903808593, 0.08825856018066407, 0.08857084655761718 ], "count": 3, "total": 0.2647500457763672, "mean": 0.08825001525878906, "p50": 0.08825856018066407, "p90": 0.08850838928222655, "p95": 0.08853961791992188, "p99": 0.08856460083007812, "stdev": 0.00026551486635231693, "stdev_": 0.3008666520608602 }, "throughput": { "unit": "samples/s", "value": 67.98865679972155 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "model": "microsoft/deberta-v3-base", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1140.764672, "max_global_vram": 4597.481472, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 5, "total": 1.0522736740112304, "mean": 0.21045473480224608, "stdev": 0.2521626361507063, "p50": 0.08427519989013672, "p90": 0.46303764648437507, "p95": 0.5889081359863281, "p99": 0.6896045275878907, "values": [ 0.7147786254882813, 0.08542617797851562, 0.08360550689697266, 0.08418816375732421, 0.08427519989013672 ] }, "throughput": { "unit": "samples/s", "value": 47.51615595342393 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1140.764672, "max_global_vram": 4597.481472, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 2, "total": 0.8002048034667969, "mean": 0.40010240173339845, "stdev": 0.3146762237548828, "p50": 0.40010240173339845, "p90": 0.6518433807373047, "p95": 0.683311003112793, "p99": 0.7084851010131836, "values": [ 0.7147786254882813, 0.08542617797851562 ] }, "throughput": { "unit": "samples/s", "value": 9.997440611879489 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1140.764672, "max_global_vram": 4597.481472, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 3, "total": 0.2520688705444336, "mean": 0.08402295684814454, "stdev": 0.0002973125946472148, "p50": 0.08418816375732421, "p90": 0.08425779266357422, "p95": 0.08426649627685547, "p99": 0.08427345916748047, "values": [ 0.08360550689697266, 0.08418816375732421, 0.08427519989013672 ] }, "throughput": { "unit": "samples/s", "value": 71.40905563278207 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null