The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code:   DatasetGenerationError
Exception:    ArrowNotImplementedError
Message:      Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
Traceback:    Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2013, in _prepare_split_single
                  writer.write_table(table)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 583, in write_table
                  self._build_writer(inferred_schema=pa_table.schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 404, in _build_writer
                  self.pa_writer = self._WRITER_CLASS(self.stream, schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__
                  self.writer = _parquet.ParquetWriter(
                File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
              
              During handling of the above exception, another exception occurred:
              
              Traceback (most recent call last):
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2029, in _prepare_split_single
                  num_examples, num_bytes = writer.finalize()
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 602, in finalize
                  self._build_writer(self.schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 404, in _build_writer
                  self.pa_writer = self._WRITER_CLASS(self.stream, schema)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__
                  self.writer = _parquet.ParquetWriter(
                File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__
                File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status
                File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status
              pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field.
              
              The above exception was the direct cause of the following exception:
              
              Traceback (most recent call last):
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1396, in compute_config_parquet_and_info_response
                  parquet_operations = convert_to_parquet(builder)
                File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1045, in convert_to_parquet
                  builder.download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1029, in download_and_prepare
                  self._download_and_prepare(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1124, in _download_and_prepare
                  self._prepare_split(split_generator, **prepare_split_kwargs)
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1884, in _prepare_split
                  for job_id, done, content in self._prepare_split_single(
                File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2040, in _prepare_split_single
                  raise DatasetGenerationError("An error occurred while generating the dataset") from e
              datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

config
dict
report
dict
name
string
backend
dict
scenario
dict
launcher
dict
environment
dict
overall
dict
warmup
dict
train
dict
{ "name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "library": "transformers", "model_type": "bert", "model": "google-bert/bert-base-uncased", "processor": "google-bert/bert-base-uncased", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1290.678272, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 5, "total": 0.5068982505798341, "mean": 0.10137965011596681, "stdev": 0.11641608256321201, "p50": 0.04309702301025391, "p90": 0.21814531250000002, "p95": 0.2761776214599609, "p99": 0.32260346862792966, "values": [ 0.33420993041992186, 0.04404838562011719, 0.04309702301025391, 0.042807296752929686, 0.04273561477661133 ] }, "throughput": { "unit": "samples/s", "value": 98.63912519486047 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1290.678272, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 2, "total": 0.37825831604003907, "mean": 0.18912915802001953, "stdev": 0.14508077239990233, "p50": 0.18912915802001953, "p90": 0.3051937759399414, "p95": 0.3197018531799316, "p99": 0.3313083149719238, "values": [ 0.33420993041992186, 0.04404838562011719 ] }, "throughput": { "unit": "samples/s", "value": 21.14956806171894 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1290.678272, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 3, "total": 0.12863993453979491, "mean": 0.04287997817993164, "stdev": 0.00015623896272364855, "p50": 0.042807296752929686, "p90": 0.04303907775878906, "p95": 0.04306805038452149, "p99": 0.04309122848510742, "values": [ 0.04309702301025391, 0.042807296752929686, 0.04273561477661133 ] }, "throughput": { "unit": "samples/s", "value": 139.92544433728455 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_fill-mask_google-bert/bert-base-uncased
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "library": "transformers", "model_type": "bert", "model": "google-bert/bert-base-uncased", "processor": "google-bert/bert-base-uncased", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null }
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1290.678272, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 5, "total": 0.5068982505798341, "mean": 0.10137965011596681, "stdev": 0.11641608256321201, "p50": 0.04309702301025391, "p90": 0.21814531250000002, "p95": 0.2761776214599609, "p99": 0.32260346862792966, "values": [ 0.33420993041992186, 0.04404838562011719, 0.04309702301025391, 0.042807296752929686, 0.04273561477661133 ] }, "throughput": { "unit": "samples/s", "value": 98.63912519486047 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1290.678272, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 2, "total": 0.37825831604003907, "mean": 0.18912915802001953, "stdev": 0.14508077239990233, "p50": 0.18912915802001953, "p90": 0.3051937759399414, "p95": 0.3197018531799316, "p99": 0.3313083149719238, "values": [ 0.33420993041992186, 0.04404838562011719 ] }, "throughput": { "unit": "samples/s", "value": 21.14956806171894 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1290.678272, "max_global_vram": 3176.660992, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 3, "total": 0.12863993453979491, "mean": 0.04287997817993164, "stdev": 0.00015623896272364855, "p50": 0.042807296752929686, "p90": 0.04303907775878906, "p95": 0.04306805038452149, "p99": 0.04309122848510742, "values": [ 0.04309702301025391, 0.042807296752929686, 0.04273561477661133 ] }, "throughput": { "unit": "samples/s", "value": 139.92544433728455 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "fill-mask", "model": "google-bert/bert-base-uncased", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1063.8336, "max_global_vram": 3169.32096, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 5, "total": 0.7448790740966797, "mean": 0.14897581481933594, "stdev": 0.2054173633207176, "p50": 0.04632883071899414, "p90": 0.35471870422363283, "p95": 0.4572641067504882, "p99": 0.5393004287719726, "values": [ 0.5598095092773437, 0.04708249664306641, 0.04632883071899414, 0.04576665496826172, 0.04589158248901367 ] }, "throughput": { "unit": "samples/s", "value": 67.1249894630687 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1063.8336, "max_global_vram": 3169.32096, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 2, "total": 0.6068920059204101, "mean": 0.3034460029602051, "stdev": 0.25636350631713867, "p50": 0.3034460029602051, "p90": 0.508536808013916, "p95": 0.5341731586456299, "p99": 0.554682239151001, "values": [ 0.5598095092773437, 0.04708249664306641 ] }, "throughput": { "unit": "samples/s", "value": 13.181916917602548 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1063.8336, "max_global_vram": 3169.32096, "max_process_vram": 0, "max_reserved": 2520.776704, "max_allocated": 2211.86048 }, "latency": { "unit": "s", "count": 3, "total": 0.13798706817626955, "mean": 0.045995689392089846, "stdev": 0.00024102431292157434, "p50": 0.04589158248901367, "p90": 0.04624138107299805, "p95": 0.0462851058959961, "p99": 0.04632008575439454, "values": [ 0.04632883071899414, 0.04576665496826172, 0.04589158248901367 ] }, "throughput": { "unit": "samples/s", "value": 130.4470066499722 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "library": "transformers", "model_type": "vit", "model": "google/vit-base-patch16-224", "processor": "google/vit-base-patch16-224", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1621.05344, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "count": 5, "total": 0.4880517120361328, "mean": 0.09761034240722656, "stdev": 0.1142101078717083, "p50": 0.04050636672973633, "p90": 0.21193584594726564, "p95": 0.26898309020996086, "p99": 0.31462088562011714, "values": [ 0.3260303344726562, 0.04079411315917969, 0.04050636672973633, 0.04039475250244141, 0.04032614517211914 ] }, "throughput": { "unit": "samples/s", "value": 102.44816023982774 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1621.05344, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "count": 2, "total": 0.3668244476318359, "mean": 0.18341222381591796, "stdev": 0.14261811065673827, "p50": 0.18341222381591796, "p90": 0.2975067123413086, "p95": 0.3117685234069824, "p99": 0.32317797225952144, "values": [ 0.3260303344726562, 0.04079411315917969 ] }, "throughput": { "unit": "samples/s", "value": 21.808797237062063 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1621.05344, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "count": 3, "total": 0.12122726440429688, "mean": 0.040409088134765625, "stdev": 0.00007427016100289037, "p50": 0.04039475250244141, "p90": 0.04048404388427734, "p95": 0.040495205307006836, "p99": 0.04050413444519043, "values": [ 0.04050636672973633, 0.04039475250244141, 0.04032614517211914 ] }, "throughput": { "unit": "samples/s", "value": 148.4814500141603 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_image-classification_google/vit-base-patch16-224
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "library": "transformers", "model_type": "vit", "model": "google/vit-base-patch16-224", "processor": "google/vit-base-patch16-224", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null }
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1621.05344, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "count": 5, "total": 0.4880517120361328, "mean": 0.09761034240722656, "stdev": 0.1142101078717083, "p50": 0.04050636672973633, "p90": 0.21193584594726564, "p95": 0.26898309020996086, "p99": 0.31462088562011714, "values": [ 0.3260303344726562, 0.04079411315917969, 0.04050636672973633, 0.04039475250244141, 0.04032614517211914 ] }, "throughput": { "unit": "samples/s", "value": 102.44816023982774 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1621.05344, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "count": 2, "total": 0.3668244476318359, "mean": 0.18341222381591796, "stdev": 0.14261811065673827, "p50": 0.18341222381591796, "p90": 0.2975067123413086, "p95": 0.3117685234069824, "p99": 0.32317797225952144, "values": [ 0.3260303344726562, 0.04079411315917969 ] }, "throughput": { "unit": "samples/s", "value": 21.808797237062063 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1621.05344, "max_global_vram": 2618.81856, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1755.291648 }, "latency": { "unit": "s", "count": 3, "total": 0.12122726440429688, "mean": 0.040409088134765625, "stdev": 0.00007427016100289037, "p50": 0.04039475250244141, "p90": 0.04048404388427734, "p95": 0.040495205307006836, "p99": 0.04050413444519043, "values": [ 0.04050636672973633, 0.04039475250244141, 0.04032614517211914 ] }, "throughput": { "unit": "samples/s", "value": 148.4814500141603 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "image-classification", "model": "google/vit-base-patch16-224", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1446.354944, "max_global_vram": 2628.255744, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1756.126208 }, "latency": { "unit": "s", "count": 5, "total": 0.48406525039672854, "mean": 0.09681305007934571, "stdev": 0.1110534407118009, "p50": 0.04146995162963867, "p90": 0.20794796142578126, "p95": 0.26343380432128904, "p99": 0.3078224786376953, "values": [ 0.3189196472167969, 0.04103168106079102, 0.041490432739257815, 0.04146995162963867, 0.04115353775024414 ] }, "throughput": { "unit": "samples/s", "value": 103.29185984538483 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1446.354944, "max_global_vram": 2628.255744, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1756.126208 }, "latency": { "unit": "s", "count": 2, "total": 0.3599513282775879, "mean": 0.17997566413879396, "stdev": 0.13894398307800293, "p50": 0.17997566413879396, "p90": 0.2911308506011963, "p95": 0.30502524890899657, "p99": 0.31614076755523685, "values": [ 0.3189196472167969, 0.04103168106079102 ] }, "throughput": { "unit": "samples/s", "value": 22.225227055782792 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1446.354944, "max_global_vram": 2628.255744, "max_process_vram": 0, "max_reserved": 1956.642816, "max_allocated": 1756.126208 }, "latency": { "unit": "s", "count": 3, "total": 0.12411392211914063, "mean": 0.041371307373046874, "stdev": 0.00015421321911462263, "p50": 0.04146995162963867, "p90": 0.04148633651733399, "p95": 0.0414883846282959, "p99": 0.041490023117065435, "values": [ 0.041490432739257815, 0.04146995162963867, 0.04115353775024414 ] }, "throughput": { "unit": "samples/s", "value": 145.02804917180256 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1308.54912, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 5, "total": 0.5452933044433593, "mean": 0.10905866088867186, "stdev": 0.11705672773476211, "p50": 0.0504453125, "p90": 0.22648237762451173, "p95": 0.28482621078491205, "p99": 0.33150127731323237, "values": [ 0.3431700439453125, 0.05145087814331055, 0.050187263488769535, 0.0500398063659668, 0.0504453125 ] }, "throughput": { "unit": "samples/s", "value": 91.69377212698492 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1308.54912, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 2, "total": 0.39462092208862304, "mean": 0.19731046104431152, "stdev": 0.14585958290100096, "p50": 0.19731046104431152, "p90": 0.3139981273651123, "p95": 0.32858408565521235, "p99": 0.34025285228729246, "values": [ 0.3431700439453125, 0.05145087814331055 ] }, "throughput": { "unit": "samples/s", "value": 20.27262000620276 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1308.54912, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 3, "total": 0.15067238235473635, "mean": 0.05022412745157878, "stdev": 0.00016758683321418588, "p50": 0.050187263488769535, "p90": 0.05039370269775391, "p95": 0.05041950759887695, "p99": 0.05044015151977539, "values": [ 0.050187263488769535, 0.0500398063659668, 0.0504453125 ] }, "throughput": { "unit": "samples/s", "value": 119.46449454566665 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_multiple-choice_FacebookAI/roberta-base
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null }
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1308.54912, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 5, "total": 0.5452933044433593, "mean": 0.10905866088867186, "stdev": 0.11705672773476211, "p50": 0.0504453125, "p90": 0.22648237762451173, "p95": 0.28482621078491205, "p99": 0.33150127731323237, "values": [ 0.3431700439453125, 0.05145087814331055, 0.050187263488769535, 0.0500398063659668, 0.0504453125 ] }, "throughput": { "unit": "samples/s", "value": 91.69377212698492 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1308.54912, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 2, "total": 0.39462092208862304, "mean": 0.19731046104431152, "stdev": 0.14585958290100096, "p50": 0.19731046104431152, "p90": 0.3139981273651123, "p95": 0.32858408565521235, "p99": 0.34025285228729246, "values": [ 0.3431700439453125, 0.05145087814331055 ] }, "throughput": { "unit": "samples/s", "value": 20.27262000620276 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1308.54912, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 3, "total": 0.15067238235473635, "mean": 0.05022412745157878, "stdev": 0.00016758683321418588, "p50": 0.050187263488769535, "p90": 0.05039370269775391, "p95": 0.05041950759887695, "p99": 0.05044015151977539, "values": [ 0.050187263488769535, 0.0500398063659668, 0.0504453125 ] }, "throughput": { "unit": "samples/s", "value": 119.46449454566665 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "multiple-choice", "model": "FacebookAI/roberta-base", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1093.496832, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 5, "total": 0.8026234703063965, "mean": 0.16052469406127928, "stdev": 0.22240891148008993, "p50": 0.04907724761962891, "p90": 0.38326721343994147, "p95": 0.49430444412231433, "p99": 0.5831342286682129, "values": [ 0.6053416748046875, 0.05015552139282226, 0.04897484970092773, 0.04907417678833008, 0.04907724761962891 ] }, "throughput": { "unit": "samples/s", "value": 62.295711313940046 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1093.496832, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 2, "total": 0.6554971961975098, "mean": 0.3277485980987549, "stdev": 0.27759307670593264, "p50": 0.3277485980987549, "p90": 0.549823059463501, "p95": 0.5775823671340942, "p99": 0.5997898132705688, "values": [ 0.6053416748046875, 0.05015552139282226 ] }, "throughput": { "unit": "samples/s", "value": 12.204476306546239 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1093.496832, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.23424 }, "latency": { "unit": "s", "count": 3, "total": 0.14712627410888673, "mean": 0.049042091369628914, "stdev": 0.000047563564546161045, "p50": 0.04907417678833008, "p90": 0.04907663345336914, "p95": 0.049076940536499025, "p99": 0.04907718620300293, "values": [ 0.04897484970092773, 0.04907417678833008, 0.04907724761962891 ] }, "throughput": { "unit": "samples/s", "value": 122.34388527149387 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1295.794176, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 5, "total": 0.547629940032959, "mean": 0.10952598800659179, "stdev": 0.11489268215249583, "p50": 0.05175603103637695, "p90": 0.22485728607177735, "p95": 0.2820829093933105, "p99": 0.3278634080505371, "values": [ 0.33930853271484374, 0.05318041610717773, 0.05175603103637695, 0.0517344970703125, 0.05165046310424805 ] }, "throughput": { "unit": "samples/s", "value": 91.3025317735381 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1295.794176, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 2, "total": 0.39248894882202146, "mean": 0.19624447441101073, "stdev": 0.143064058303833, "p50": 0.19624447441101073, "p90": 0.31069572105407717, "p95": 0.3250021268844604, "p99": 0.3364472515487671, "values": [ 0.33930853271484374, 0.05318041610717773 ] }, "throughput": { "unit": "samples/s", "value": 20.382739498807368 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1295.794176, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 3, "total": 0.1551409912109375, "mean": 0.051713663736979165, "stdev": 0.00004554607854928572, "p50": 0.0517344970703125, "p90": 0.051751724243164064, "p95": 0.051753877639770504, "p99": 0.05175560035705566, "values": [ 0.05175603103637695, 0.0517344970703125, 0.05165046310424805 ] }, "throughput": { "unit": "samples/s", "value": 116.02349488360748 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_text-classification_FacebookAI/roberta-base
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null }
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1295.794176, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 5, "total": 0.547629940032959, "mean": 0.10952598800659179, "stdev": 0.11489268215249583, "p50": 0.05175603103637695, "p90": 0.22485728607177735, "p95": 0.2820829093933105, "p99": 0.3278634080505371, "values": [ 0.33930853271484374, 0.05318041610717773, 0.05175603103637695, 0.0517344970703125, 0.05165046310424805 ] }, "throughput": { "unit": "samples/s", "value": 91.3025317735381 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1295.794176, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 2, "total": 0.39248894882202146, "mean": 0.19624447441101073, "stdev": 0.143064058303833, "p50": 0.19624447441101073, "p90": 0.31069572105407717, "p95": 0.3250021268844604, "p99": 0.3364472515487671, "values": [ 0.33930853271484374, 0.05318041610717773 ] }, "throughput": { "unit": "samples/s", "value": 20.382739498807368 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1295.794176, "max_global_vram": 3386.376192, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 3, "total": 0.1551409912109375, "mean": 0.051713663736979165, "stdev": 0.00004554607854928572, "p50": 0.0517344970703125, "p90": 0.051751724243164064, "p95": 0.051753877639770504, "p99": 0.05175560035705566, "values": [ 0.05175603103637695, 0.0517344970703125, 0.05165046310424805 ] }, "throughput": { "unit": "samples/s", "value": 116.02349488360748 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "model": "FacebookAI/roberta-base", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1080.1152, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 5, "total": 0.778464241027832, "mean": 0.1556928482055664, "stdev": 0.21129125718980005, "p50": 0.05020159912109375, "p90": 0.36730080566406254, "p95": 0.4727875488281249, "p99": 0.557176943359375, "values": [ 0.5782742919921875, 0.050840576171875, 0.04969574356079102, 0.049452030181884765, 0.05020159912109375 ] }, "throughput": { "unit": "samples/s", "value": 64.22902602948513 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1080.1152, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 2, "total": 0.6291148681640625, "mean": 0.31455743408203124, "stdev": 0.26371685791015625, "p50": 0.31455743408203124, "p90": 0.5255309204101563, "p95": 0.5519026062011718, "p99": 0.5729999548339844, "values": [ 0.5782742919921875, 0.050840576171875 ] }, "throughput": { "unit": "samples/s", "value": 12.716278703357135 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1080.1152, "max_global_vram": 3379.03616, "max_process_vram": 0, "max_reserved": 2730.491904, "max_allocated": 2516.250112 }, "latency": { "unit": "s", "count": 3, "total": 0.14934937286376954, "mean": 0.049783124287923176, "stdev": 0.0003121857804388603, "p50": 0.04969574356079102, "p90": 0.05010042800903321, "p95": 0.050151013565063476, "p99": 0.0501914820098877, "values": [ 0.04969574356079102, 0.049452030181884765, 0.05020159912109375 ] }, "throughput": { "unit": "samples/s", "value": 120.52276922795568 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_text-generation_openai-community/gpt2", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "library": "transformers", "model_type": "gpt2", "model": "openai-community/gpt2", "processor": "openai-community/gpt2", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1340.739584, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 5, "total": 0.5288140792846681, "mean": 0.10576281585693362, "stdev": 0.12384743582306901, "p50": 0.043684864044189455, "p90": 0.22980915222167975, "p95": 0.29163315277099605, "p99": 0.34109235321044923, "values": [ 0.3534571533203125, 0.04433715057373047, 0.043670528411865236, 0.04366438293457031, 0.043684864044189455 ] }, "throughput": { "unit": "samples/s", "value": 94.55118908262708 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1340.739584, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 2, "total": 0.397794303894043, "mean": 0.1988971519470215, "stdev": 0.15456000137329104, "p50": 0.1988971519470215, "p90": 0.32254515304565434, "p95": 0.3380011531829834, "p99": 0.3503659532928467, "values": [ 0.3534571533203125, 0.04433715057373047 ] }, "throughput": { "unit": "samples/s", "value": 20.110896314218945 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1340.739584, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 3, "total": 0.131019775390625, "mean": 0.04367325846354167, "stdev": 0.00000858133048832658, "p50": 0.043670528411865236, "p90": 0.04368199691772461, "p95": 0.04368343048095703, "p99": 0.04368457733154297, "values": [ 0.043670528411865236, 0.04366438293457031, 0.043684864044189455 ] }, "throughput": { "unit": "samples/s", "value": 137.38384107539824 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_text-generation_openai-community/gpt2
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "library": "transformers", "model_type": "gpt2", "model": "openai-community/gpt2", "processor": "openai-community/gpt2", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null }
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1340.739584, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 5, "total": 0.5288140792846681, "mean": 0.10576281585693362, "stdev": 0.12384743582306901, "p50": 0.043684864044189455, "p90": 0.22980915222167975, "p95": 0.29163315277099605, "p99": 0.34109235321044923, "values": [ 0.3534571533203125, 0.04433715057373047, 0.043670528411865236, 0.04366438293457031, 0.043684864044189455 ] }, "throughput": { "unit": "samples/s", "value": 94.55118908262708 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1340.739584, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 2, "total": 0.397794303894043, "mean": 0.1988971519470215, "stdev": 0.15456000137329104, "p50": 0.1988971519470215, "p90": 0.32254515304565434, "p95": 0.3380011531829834, "p99": 0.3503659532928467, "values": [ 0.3534571533203125, 0.04433715057373047 ] }, "throughput": { "unit": "samples/s", "value": 20.110896314218945 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1340.739584, "max_global_vram": 3566.731264, "max_process_vram": 0, "max_reserved": 2910.846976, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 3, "total": 0.131019775390625, "mean": 0.04367325846354167, "stdev": 0.00000858133048832658, "p50": 0.043670528411865236, "p90": 0.04368199691772461, "p95": 0.04368343048095703, "p99": 0.04368457733154297, "values": [ 0.043670528411865236, 0.04366438293457031, 0.043684864044189455 ] }, "throughput": { "unit": "samples/s", "value": 137.38384107539824 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_text-generation_openai-community/gpt2", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "model": "openai-community/gpt2", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1107.173376, "max_global_vram": 3563.585536, "max_process_vram": 0, "max_reserved": 2915.04128, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 5, "total": 0.8139294586181639, "mean": 0.16278589172363278, "stdev": 0.2273662362653359, "p50": 0.04927385711669922, "p90": 0.3902293930053711, "p95": 0.5038737297058105, "p99": 0.594789199066162, "values": [ 0.61751806640625, 0.049296382904052735, 0.04860006332397461, 0.0492410888671875, 0.04927385711669922 ] }, "throughput": { "unit": "samples/s", "value": 61.4303849929289 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1107.173376, "max_global_vram": 3563.585536, "max_process_vram": 0, "max_reserved": 2915.04128, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 2, "total": 0.6668144493103026, "mean": 0.3334072246551513, "stdev": 0.2841108417510986, "p50": 0.3334072246551513, "p90": 0.5606958980560303, "p95": 0.5891069822311401, "p99": 0.6118358495712279, "values": [ 0.61751806640625, 0.049296382904052735 ] }, "throughput": { "unit": "samples/s", "value": 11.99734050195603 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1107.173376, "max_global_vram": 3563.585536, "max_process_vram": 0, "max_reserved": 2915.04128, "max_allocated": 2523.776 }, "latency": { "unit": "s", "count": 3, "total": 0.1471150093078613, "mean": 0.04903833643595377, "stdev": 0.00031019448743967263, "p50": 0.0492410888671875, "p90": 0.04926730346679687, "p95": 0.049270580291748044, "p99": 0.04927320175170898, "values": [ 0.04860006332397461, 0.0492410888671875, 0.04927385711669922 ] }, "throughput": { "unit": "samples/s", "value": 122.35325331307405 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
{ "name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base", "backend": { "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "library": "transformers", "model_type": "deberta-v2", "model": "microsoft/deberta-v3-base", "processor": "microsoft/deberta-v3-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1377.017856, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 5, "total": 0.7545148620605469, "mean": 0.15090297241210937, "stdev": 0.13075304596870557, "p50": 0.0851813735961914, "p90": 0.28223302917480475, "p95": 0.34731887664794914, "p99": 0.3993875546264648, "values": [ 0.41240472412109375, 0.08697548675537109, 0.08489266967773437, 0.08506060791015625, 0.0851813735961914 ] }, "throughput": { "unit": "samples/s", "value": 66.2677470175368 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1377.017856, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 2, "total": 0.49938021087646484, "mean": 0.24969010543823242, "stdev": 0.16271461868286133, "p50": 0.24969010543823242, "p90": 0.3798618003845215, "p95": 0.3961332622528076, "p99": 0.40915043174743654, "values": [ 0.41240472412109375, 0.08697548675537109 ] }, "throughput": { "unit": "samples/s", "value": 16.019857867333506 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1377.017856, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 3, "total": 0.2551346511840821, "mean": 0.08504488372802736, "stdev": 0.00011838616349565163, "p50": 0.08506060791015625, "p90": 0.08515722045898437, "p95": 0.0851692970275879, "p99": 0.0851789582824707, "values": [ 0.08489266967773437, 0.08506060791015625, 0.0851813735961914 ] }, "throughput": { "unit": "samples/s", "value": 70.55098128169516 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null
null
null
cuda_training_transformers_token-classification_microsoft/deberta-v3-base
{ "name": "pytorch", "version": "2.4.1+cu124", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "library": "transformers", "model_type": "deberta-v2", "model": "microsoft/deberta-v3-base", "processor": "microsoft/deberta-v3-base", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }
{ "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "evaluation_strategy": "no", "eval_strategy": "no", "save_strategy": "no", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }
{ "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }
{ "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.261056, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.2", "transformers_commit": null, "accelerate_version": "0.34.2", "accelerate_commit": null, "diffusers_version": "0.30.3", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.9", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null }
null
null
null
null
null
null
null
null
null
null
{ "memory": { "unit": "MB", "max_ram": 1377.017856, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 5, "total": 0.7545148620605469, "mean": 0.15090297241210937, "stdev": 0.13075304596870557, "p50": 0.0851813735961914, "p90": 0.28223302917480475, "p95": 0.34731887664794914, "p99": 0.3993875546264648, "values": [ 0.41240472412109375, 0.08697548675537109, 0.08489266967773437, 0.08506060791015625, 0.0851813735961914 ] }, "throughput": { "unit": "samples/s", "value": 66.2677470175368 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1377.017856, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 2, "total": 0.49938021087646484, "mean": 0.24969010543823242, "stdev": 0.16271461868286133, "p50": 0.24969010543823242, "p90": 0.3798618003845215, "p95": 0.3961332622528076, "p99": 0.40915043174743654, "values": [ 0.41240472412109375, 0.08697548675537109 ] }, "throughput": { "unit": "samples/s", "value": 16.019857867333506 }, "energy": null, "efficiency": null }
{ "memory": { "unit": "MB", "max_ram": 1377.017856, "max_global_vram": 4604.821504, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 3, "total": 0.2551346511840821, "mean": 0.08504488372802736, "stdev": 0.00011838616349565163, "p50": 0.08506060791015625, "p90": 0.08515722045898437, "p95": 0.0851692970275879, "p99": 0.0851789582824707, "values": [ 0.08489266967773437, 0.08506060791015625, 0.0851813735961914 ] }, "throughput": { "unit": "samples/s", "value": 70.55098128169516 }, "energy": null, "efficiency": null }
{ "name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base", "backend": { "name": "pytorch", "version": "2.2.2", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "token-classification", "model": "microsoft/deberta-v3-base", "library": "transformers", "device": "cuda", "device_ids": "0", "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "hub_kwargs": { "revision": "main", "force_download": false, "local_files_only": false, "trust_remote_code": false }, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "training", "_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario", "max_steps": 5, "warmup_steps": 2, "dataset_shapes": { "dataset_size": 500, "sequence_length": 16, "num_choices": 1 }, "training_arguments": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 1, "output_dir": "./trainer_output", "do_train": true, "use_cpu": false, "max_steps": 5, "do_eval": false, "do_predict": false, "report_to": "none", "skip_memory_metrics": true, "ddp_find_unused_parameters": false }, "latency": true, "memory": true, "energy": false }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": true, "device_isolation_action": "error", "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7R32", "cpu_count": 16, "cpu_ram_mb": 66697.29792, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.14", "gpu": [ "NVIDIA A10G" ], "gpu_count": 1, "gpu_vram_mb": 24146608128, "optimum_benchmark_version": "0.2.0", "optimum_benchmark_commit": null, "transformers_version": "4.40.2", "transformers_commit": null, "accelerate_version": "0.30.0", "accelerate_commit": null, "diffusers_version": "0.27.2", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "0.9.16", "timm_commit": null, "peft_version": null, "peft_commit": null } }
{ "overall": { "memory": { "unit": "MB", "max_ram": 1140.764672, "max_global_vram": 4597.481472, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 5, "total": 1.0522736740112304, "mean": 0.21045473480224608, "stdev": 0.2521626361507063, "p50": 0.08427519989013672, "p90": 0.46303764648437507, "p95": 0.5889081359863281, "p99": 0.6896045275878907, "values": [ 0.7147786254882813, 0.08542617797851562, 0.08360550689697266, 0.08418816375732421, 0.08427519989013672 ] }, "throughput": { "unit": "samples/s", "value": 47.51615595342393 }, "energy": null, "efficiency": null }, "warmup": { "memory": { "unit": "MB", "max_ram": 1140.764672, "max_global_vram": 4597.481472, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 2, "total": 0.8002048034667969, "mean": 0.40010240173339845, "stdev": 0.3146762237548828, "p50": 0.40010240173339845, "p90": 0.6518433807373047, "p95": 0.683311003112793, "p99": 0.7084851010131836, "values": [ 0.7147786254882813, 0.08542617797851562 ] }, "throughput": { "unit": "samples/s", "value": 9.997440611879489 }, "energy": null, "efficiency": null }, "train": { "memory": { "unit": "MB", "max_ram": 1140.764672, "max_global_vram": 4597.481472, "max_process_vram": 0, "max_reserved": 3948.937216, "max_allocated": 3702.95552 }, "latency": { "unit": "s", "count": 3, "total": 0.2520688705444336, "mean": 0.08402295684814454, "stdev": 0.0002973125946472148, "p50": 0.08418816375732421, "p90": 0.08425779266357422, "p95": 0.08426649627685547, "p99": 0.08427345916748047, "values": [ 0.08360550689697266, 0.08418816375732421, 0.08427519989013672 ] }, "throughput": { "unit": "samples/s", "value": 71.40905563278207 }, "energy": null, "efficiency": null } }
null
null
null
null
null
null
null
null