The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code: DatasetGenerationError Exception: ArrowNotImplementedError Message: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1869, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 578, in write_table self._build_writer(inferred_schema=pa_table.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 399, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1885, in _prepare_split_single num_examples, num_bytes = writer.finalize() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 597, in finalize self._build_writer(self.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 399, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1010, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 2157, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1392, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1041, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 924, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 999, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1740, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1896, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
config
dict | report
dict | name
string | backend
dict | scenario
dict | launcher
dict | environment
dict | print_report
bool | log_report
bool | overall
dict | warmup
dict | train
dict |
---|---|---|---|---|---|---|---|---|---|---|---|
{
"name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1299.668992,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.3378821105957031,
0.043873279571533204,
0.04286873626708984,
0.04303155136108398,
0.04216320037841797
],
"count": 5,
"total": 0.5098188781738281,
"mean": 0.10196377563476562,
"p50": 0.04303155136108398,
"p90": 0.2202785781860352,
"p95": 0.2790803443908691,
"p99": 0.3261217573547363,
"stdev": 0.1179604222104298,
"stdev_": 115.68855848665724
},
"throughput": {
"unit": "samples/s",
"value": 98.07404578484827
},
"energy": {
"unit": "kWh",
"cpu": 0.00001023778893402686,
"ram": 0.000005588101379470862,
"gpu": 0.000017930014344001016,
"total": 0.00003375590465749874
},
"efficiency": {
"unit": "samples/kWh",
"value": 296244.4674928462
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1299.668992,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.3378821105957031,
0.043873279571533204
],
"count": 2,
"total": 0.3817553901672363,
"mean": 0.19087769508361815,
"p50": 0.19087769508361815,
"p90": 0.3084812274932861,
"p95": 0.3231816690444946,
"p99": 0.3349420222854614,
"stdev": 0.14700441551208496,
"stdev_": 77.01497833347499
},
"throughput": {
"unit": "samples/s",
"value": 20.95582722877973
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1299.668992,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.04286873626708984,
0.04303155136108398,
0.04216320037841797
],
"count": 3,
"total": 0.12806348800659179,
"mean": 0.04268782933553059,
"p50": 0.04286873626708984,
"p90": 0.042998988342285153,
"p95": 0.04301526985168457,
"p99": 0.0430282950592041,
"stdev": 0.0003768765011591056,
"stdev_": 0.8828663978128725
},
"throughput": {
"unit": "samples/s",
"value": 140.55528457161412
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_fill-mask_google-bert/bert-base-uncased | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model_type": "bert",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1299.668992,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.3378821105957031,
0.043873279571533204,
0.04286873626708984,
0.04303155136108398,
0.04216320037841797
],
"count": 5,
"total": 0.5098188781738281,
"mean": 0.10196377563476562,
"p50": 0.04303155136108398,
"p90": 0.2202785781860352,
"p95": 0.2790803443908691,
"p99": 0.3261217573547363,
"stdev": 0.1179604222104298,
"stdev_": 115.68855848665724
},
"throughput": {
"unit": "samples/s",
"value": 98.07404578484827
},
"energy": {
"unit": "kWh",
"cpu": 0.00001023778893402686,
"ram": 0.000005588101379470862,
"gpu": 0.000017930014344001016,
"total": 0.00003375590465749874
},
"efficiency": {
"unit": "samples/kWh",
"value": 296244.4674928462
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1299.668992,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.3378821105957031,
0.043873279571533204
],
"count": 2,
"total": 0.3817553901672363,
"mean": 0.19087769508361815,
"p50": 0.19087769508361815,
"p90": 0.3084812274932861,
"p95": 0.3231816690444946,
"p99": 0.3349420222854614,
"stdev": 0.14700441551208496,
"stdev_": 77.01497833347499
},
"throughput": {
"unit": "samples/s",
"value": 20.95582722877973
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1299.668992,
"max_global_vram": 3176.660992,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"values": [
0.04286873626708984,
0.04303155136108398,
0.04216320037841797
],
"count": 3,
"total": 0.12806348800659179,
"mean": 0.04268782933553059,
"p50": 0.04286873626708984,
"p90": 0.042998988342285153,
"p95": 0.04301526985168457,
"p99": 0.0430282950592041,
"stdev": 0.0003768765011591056,
"stdev_": 0.8828663978128725
},
"throughput": {
"unit": "samples/s",
"value": 140.55528457161412
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"model": "google-bert/bert-base-uncased",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.7448790740966797,
"mean": 0.14897581481933594,
"stdev": 0.2054173633207176,
"p50": 0.04632883071899414,
"p90": 0.35471870422363283,
"p95": 0.4572641067504882,
"p99": 0.5393004287719726,
"values": [
0.5598095092773437,
0.04708249664306641,
0.04632883071899414,
0.04576665496826172,
0.04589158248901367
]
},
"throughput": {
"unit": "samples/s",
"value": 67.1249894630687
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6068920059204101,
"mean": 0.3034460029602051,
"stdev": 0.25636350631713867,
"p50": 0.3034460029602051,
"p90": 0.508536808013916,
"p95": 0.5341731586456299,
"p99": 0.554682239151001,
"values": [
0.5598095092773437,
0.04708249664306641
]
},
"throughput": {
"unit": "samples/s",
"value": 13.181916917602548
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1063.8336,
"max_global_vram": 3169.32096,
"max_process_vram": 0,
"max_reserved": 2520.776704,
"max_allocated": 2211.86048
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.13798706817626955,
"mean": 0.045995689392089846,
"stdev": 0.00024102431292157434,
"p50": 0.04589158248901367,
"p90": 0.04624138107299805,
"p95": 0.0462851058959961,
"p99": 0.04632008575439454,
"values": [
0.04632883071899414,
0.04576665496826172,
0.04589158248901367
]
},
"throughput": {
"unit": "samples/s",
"value": 130.4470066499722
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1631.678464,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.3338004455566406,
0.04012441635131836,
0.039790592193603515,
0.03977519989013672,
0.039820289611816405
],
"count": 5,
"total": 0.49331094360351563,
"mean": 0.09866218872070312,
"p50": 0.039820289611816405,
"p90": 0.21633003387451175,
"p95": 0.2750652397155761,
"p99": 0.3220534043884277,
"stdev": 0.11756919838438508,
"stdev_": 119.1633795163461
},
"throughput": {
"unit": "samples/s",
"value": 101.3559513493908
},
"energy": {
"unit": "kWh",
"cpu": 0.000011456950427083257,
"ram": 0.000006241313718108779,
"gpu": 0.000023570574411999404,
"total": 0.00004126883855719144
},
"efficiency": {
"unit": "samples/kWh",
"value": 242313.5796792957
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1631.678464,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.3338004455566406,
0.04012441635131836
],
"count": 2,
"total": 0.373924861907959,
"mean": 0.1869624309539795,
"p50": 0.1869624309539795,
"p90": 0.3044328426361084,
"p95": 0.3191166440963745,
"p99": 0.3308636852645874,
"stdev": 0.14683801460266113,
"stdev_": 78.53878121578612
},
"throughput": {
"unit": "samples/s",
"value": 21.39467260662969
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1631.678464,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.039790592193603515,
0.03977519989013672,
0.039820289611816405
],
"count": 3,
"total": 0.11938608169555665,
"mean": 0.039795360565185546,
"p50": 0.039790592193603515,
"p90": 0.039814350128173824,
"p95": 0.039817319869995114,
"p99": 0.03981969566345215,
"stdev": 0.000018714054895472507,
"stdev_": 0.04702572015855601
},
"throughput": {
"unit": "samples/s",
"value": 150.77134406589653
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_image-classification_google/vit-base-patch16-224 | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model_type": "vit",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1631.678464,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.3338004455566406,
0.04012441635131836,
0.039790592193603515,
0.03977519989013672,
0.039820289611816405
],
"count": 5,
"total": 0.49331094360351563,
"mean": 0.09866218872070312,
"p50": 0.039820289611816405,
"p90": 0.21633003387451175,
"p95": 0.2750652397155761,
"p99": 0.3220534043884277,
"stdev": 0.11756919838438508,
"stdev_": 119.1633795163461
},
"throughput": {
"unit": "samples/s",
"value": 101.3559513493908
},
"energy": {
"unit": "kWh",
"cpu": 0.000011456950427083257,
"ram": 0.000006241313718108779,
"gpu": 0.000023570574411999404,
"total": 0.00004126883855719144
},
"efficiency": {
"unit": "samples/kWh",
"value": 242313.5796792957
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1631.678464,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.3338004455566406,
0.04012441635131836
],
"count": 2,
"total": 0.373924861907959,
"mean": 0.1869624309539795,
"p50": 0.1869624309539795,
"p90": 0.3044328426361084,
"p95": 0.3191166440963745,
"p99": 0.3308636852645874,
"stdev": 0.14683801460266113,
"stdev_": 78.53878121578612
},
"throughput": {
"unit": "samples/s",
"value": 21.39467260662969
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1631.678464,
"max_global_vram": 2618.81856,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1755.291648
},
"latency": {
"unit": "s",
"values": [
0.039790592193603515,
0.03977519989013672,
0.039820289611816405
],
"count": 3,
"total": 0.11938608169555665,
"mean": 0.039795360565185546,
"p50": 0.039790592193603515,
"p90": 0.039814350128173824,
"p95": 0.039817319869995114,
"p99": 0.03981969566345215,
"stdev": 0.000018714054895472507,
"stdev_": 0.04702572015855601
},
"throughput": {
"unit": "samples/s",
"value": 150.77134406589653
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"model": "google/vit-base-patch16-224",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.48406525039672854,
"mean": 0.09681305007934571,
"stdev": 0.1110534407118009,
"p50": 0.04146995162963867,
"p90": 0.20794796142578126,
"p95": 0.26343380432128904,
"p99": 0.3078224786376953,
"values": [
0.3189196472167969,
0.04103168106079102,
0.041490432739257815,
0.04146995162963867,
0.04115353775024414
]
},
"throughput": {
"unit": "samples/s",
"value": 103.29185984538483
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.3599513282775879,
"mean": 0.17997566413879396,
"stdev": 0.13894398307800293,
"p50": 0.17997566413879396,
"p90": 0.2911308506011963,
"p95": 0.30502524890899657,
"p99": 0.31614076755523685,
"values": [
0.3189196472167969,
0.04103168106079102
]
},
"throughput": {
"unit": "samples/s",
"value": 22.225227055782792
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1446.354944,
"max_global_vram": 2628.255744,
"max_process_vram": 0,
"max_reserved": 1956.642816,
"max_allocated": 1756.126208
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.12411392211914063,
"mean": 0.041371307373046874,
"stdev": 0.00015421321911462263,
"p50": 0.04146995162963867,
"p90": 0.04148633651733399,
"p95": 0.0414883846282959,
"p99": 0.041490023117065435,
"values": [
0.041490432739257815,
0.04146995162963867,
0.04115353775024414
]
},
"throughput": {
"unit": "samples/s",
"value": 145.02804917180256
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1337.192448,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.35576217651367187,
0.04703129577636719,
0.04605235290527344,
0.046074878692626955,
0.04605542373657227
],
"count": 5,
"total": 0.5409761276245116,
"mean": 0.10819522552490232,
"p50": 0.046074878692626955,
"p90": 0.23226982421875003,
"p95": 0.2940160003662109,
"p99": 0.34341294128417965,
"stdev": 0.12378404630519566,
"stdev_": 114.40804869592456
},
"throughput": {
"unit": "samples/s",
"value": 92.42552017878451
},
"energy": {
"unit": "kWh",
"cpu": 0.00001055118112430383,
"ram": 0.0000057624615054751955,
"gpu": 0.000017052791419999887,
"total": 0.00003336643404977891
},
"efficiency": {
"unit": "samples/kWh",
"value": 299702.389086024
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1337.192448,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.35576217651367187,
0.04703129577636719
],
"count": 2,
"total": 0.40279347229003903,
"mean": 0.20139673614501952,
"p50": 0.20139673614501952,
"p90": 0.3248890884399414,
"p95": 0.34032563247680664,
"p99": 0.3526748677062988,
"stdev": 0.15436544036865232,
"stdev_": 76.64743894235635
},
"throughput": {
"unit": "samples/s",
"value": 19.86129505653818
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1337.192448,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.04605235290527344,
0.046074878692626955,
0.04605542373657227
],
"count": 3,
"total": 0.13818265533447266,
"mean": 0.04606088511149089,
"p50": 0.04605542373657227,
"p90": 0.046070987701416016,
"p95": 0.046072933197021486,
"p99": 0.04607448959350586,
"stdev": 0.000009974057548636764,
"stdev_": 0.021654072700718726
},
"throughput": {
"unit": "samples/s",
"value": 130.2623687208123
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_multiple-choice_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1337.192448,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.35576217651367187,
0.04703129577636719,
0.04605235290527344,
0.046074878692626955,
0.04605542373657227
],
"count": 5,
"total": 0.5409761276245116,
"mean": 0.10819522552490232,
"p50": 0.046074878692626955,
"p90": 0.23226982421875003,
"p95": 0.2940160003662109,
"p99": 0.34341294128417965,
"stdev": 0.12378404630519566,
"stdev_": 114.40804869592456
},
"throughput": {
"unit": "samples/s",
"value": 92.42552017878451
},
"energy": {
"unit": "kWh",
"cpu": 0.00001055118112430383,
"ram": 0.0000057624615054751955,
"gpu": 0.000017052791419999887,
"total": 0.00003336643404977891
},
"efficiency": {
"unit": "samples/kWh",
"value": 299702.389086024
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1337.192448,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.35576217651367187,
0.04703129577636719
],
"count": 2,
"total": 0.40279347229003903,
"mean": 0.20139673614501952,
"p50": 0.20139673614501952,
"p90": 0.3248890884399414,
"p95": 0.34032563247680664,
"p99": 0.3526748677062988,
"stdev": 0.15436544036865232,
"stdev_": 76.64743894235635
},
"throughput": {
"unit": "samples/s",
"value": 19.86129505653818
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1337.192448,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"values": [
0.04605235290527344,
0.046074878692626955,
0.04605542373657227
],
"count": 3,
"total": 0.13818265533447266,
"mean": 0.04606088511149089,
"p50": 0.04605542373657227,
"p90": 0.046070987701416016,
"p95": 0.046072933197021486,
"p99": 0.04607448959350586,
"stdev": 0.000009974057548636764,
"stdev_": 0.021654072700718726
},
"throughput": {
"unit": "samples/s",
"value": 130.2623687208123
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.8026234703063965,
"mean": 0.16052469406127928,
"stdev": 0.22240891148008993,
"p50": 0.04907724761962891,
"p90": 0.38326721343994147,
"p95": 0.49430444412231433,
"p99": 0.5831342286682129,
"values": [
0.6053416748046875,
0.05015552139282226,
0.04897484970092773,
0.04907417678833008,
0.04907724761962891
]
},
"throughput": {
"unit": "samples/s",
"value": 62.295711313940046
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6554971961975098,
"mean": 0.3277485980987549,
"stdev": 0.27759307670593264,
"p50": 0.3277485980987549,
"p90": 0.549823059463501,
"p95": 0.5775823671340942,
"p99": 0.5997898132705688,
"values": [
0.6053416748046875,
0.05015552139282226
]
},
"throughput": {
"unit": "samples/s",
"value": 12.204476306546239
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1093.496832,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.23424
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.14712627410888673,
"mean": 0.049042091369628914,
"stdev": 0.000047563564546161045,
"p50": 0.04907417678833008,
"p90": 0.04907663345336914,
"p95": 0.049076940536499025,
"p99": 0.04907718620300293,
"values": [
0.04897484970092773,
0.04907417678833008,
0.04907724761962891
]
},
"throughput": {
"unit": "samples/s",
"value": 122.34388527149387
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1323.085824,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.3465441284179688,
0.04596223831176758,
0.045655040740966796,
0.04518291091918945,
0.045499393463134766
],
"count": 5,
"total": 0.5288437118530275,
"mean": 0.10576874237060549,
"p50": 0.045655040740966796,
"p90": 0.22631137237548832,
"p95": 0.2864277503967285,
"p99": 0.33452085281372074,
"stdev": 0.12038795535102645,
"stdev_": 113.82186518697213
},
"throughput": {
"unit": "samples/s",
"value": 94.54589111933254
},
"energy": {
"unit": "kWh",
"cpu": 0.000010490307170139897,
"ram": 0.000005714536570311505,
"gpu": 0.000018112236711999644,
"total": 0.00003431708045245105
},
"efficiency": {
"unit": "samples/kWh",
"value": 291400.08031440113
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1323.085824,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.3465441284179688,
0.04596223831176758
],
"count": 2,
"total": 0.3925063667297364,
"mean": 0.1962531833648682,
"p50": 0.1962531833648682,
"p90": 0.31648593940734865,
"p95": 0.3315150339126587,
"p99": 0.3435383095169068,
"stdev": 0.1502909450531006,
"stdev_": 76.58013107164945
},
"throughput": {
"unit": "samples/s",
"value": 20.38183499201293
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1323.085824,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.045655040740966796,
0.04518291091918945,
0.045499393463134766
],
"count": 3,
"total": 0.13633734512329101,
"mean": 0.04544578170776367,
"p50": 0.045499393463134766,
"p90": 0.04562391128540039,
"p95": 0.04563947601318359,
"p99": 0.045651927795410156,
"stdev": 0.00019643880707632385,
"stdev_": 0.4322487141700227
},
"throughput": {
"unit": "samples/s",
"value": 132.02545482840705
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_text-classification_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model_type": "roberta",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1323.085824,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.3465441284179688,
0.04596223831176758,
0.045655040740966796,
0.04518291091918945,
0.045499393463134766
],
"count": 5,
"total": 0.5288437118530275,
"mean": 0.10576874237060549,
"p50": 0.045655040740966796,
"p90": 0.22631137237548832,
"p95": 0.2864277503967285,
"p99": 0.33452085281372074,
"stdev": 0.12038795535102645,
"stdev_": 113.82186518697213
},
"throughput": {
"unit": "samples/s",
"value": 94.54589111933254
},
"energy": {
"unit": "kWh",
"cpu": 0.000010490307170139897,
"ram": 0.000005714536570311505,
"gpu": 0.000018112236711999644,
"total": 0.00003431708045245105
},
"efficiency": {
"unit": "samples/kWh",
"value": 291400.08031440113
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1323.085824,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.3465441284179688,
0.04596223831176758
],
"count": 2,
"total": 0.3925063667297364,
"mean": 0.1962531833648682,
"p50": 0.1962531833648682,
"p90": 0.31648593940734865,
"p95": 0.3315150339126587,
"p99": 0.3435383095169068,
"stdev": 0.1502909450531006,
"stdev_": 76.58013107164945
},
"throughput": {
"unit": "samples/s",
"value": 20.38183499201293
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1323.085824,
"max_global_vram": 3384.27904,
"max_process_vram": 0,
"max_reserved": 2728.394752,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"values": [
0.045655040740966796,
0.04518291091918945,
0.045499393463134766
],
"count": 3,
"total": 0.13633734512329101,
"mean": 0.04544578170776367,
"p50": 0.045499393463134766,
"p90": 0.04562391128540039,
"p95": 0.04563947601318359,
"p99": 0.045651927795410156,
"stdev": 0.00019643880707632385,
"stdev_": 0.4322487141700227
},
"throughput": {
"unit": "samples/s",
"value": 132.02545482840705
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.778464241027832,
"mean": 0.1556928482055664,
"stdev": 0.21129125718980005,
"p50": 0.05020159912109375,
"p90": 0.36730080566406254,
"p95": 0.4727875488281249,
"p99": 0.557176943359375,
"values": [
0.5782742919921875,
0.050840576171875,
0.04969574356079102,
0.049452030181884765,
0.05020159912109375
]
},
"throughput": {
"unit": "samples/s",
"value": 64.22902602948513
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6291148681640625,
"mean": 0.31455743408203124,
"stdev": 0.26371685791015625,
"p50": 0.31455743408203124,
"p90": 0.5255309204101563,
"p95": 0.5519026062011718,
"p99": 0.5729999548339844,
"values": [
0.5782742919921875,
0.050840576171875
]
},
"throughput": {
"unit": "samples/s",
"value": 12.716278703357135
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1080.1152,
"max_global_vram": 3379.03616,
"max_process_vram": 0,
"max_reserved": 2730.491904,
"max_allocated": 2516.250112
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.14934937286376954,
"mean": 0.049783124287923176,
"stdev": 0.0003121857804388603,
"p50": 0.04969574356079102,
"p90": 0.05010042800903321,
"p95": 0.050151013565063476,
"p99": 0.0501914820098877,
"values": [
0.04969574356079102,
0.049452030181884765,
0.05020159912109375
]
},
"throughput": {
"unit": "samples/s",
"value": 120.52276922795568
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1350.053888,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.35609701538085936,
0.044365825653076174,
0.04398284912109375,
0.043514881134033206,
0.04412108612060547
],
"count": 5,
"total": 0.532081657409668,
"mean": 0.1064163314819336,
"p50": 0.04412108612060547,
"p90": 0.2314045394897461,
"p95": 0.29375077743530265,
"p99": 0.34362776779174803,
"stdev": 0.12484064959463413,
"stdev_": 117.3134309895173
},
"throughput": {
"unit": "samples/s",
"value": 93.97053873913808
},
"energy": {
"unit": "kWh",
"cpu": 0.00001046066729236246,
"ram": 0.000005713810620613953,
"gpu": 0.000015273067774000063,
"total": 0.00003144754568697648
},
"efficiency": {
"unit": "samples/kWh",
"value": 317989.8393196817
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1350.053888,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.35609701538085936,
0.044365825653076174
],
"count": 2,
"total": 0.40046284103393553,
"mean": 0.20023142051696777,
"p50": 0.20023142051696777,
"p90": 0.32492389640808106,
"p95": 0.3405104558944702,
"p99": 0.35297970348358154,
"stdev": 0.1558655948638916,
"stdev_": 77.84272541315933
},
"throughput": {
"unit": "samples/s",
"value": 19.976884695082294
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1350.053888,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.04398284912109375,
0.043514881134033206,
0.04412108612060547
],
"count": 3,
"total": 0.13161881637573242,
"mean": 0.04387293879191081,
"p50": 0.04398284912109375,
"p90": 0.04409343872070312,
"p95": 0.0441072624206543,
"p99": 0.044118321380615236,
"stdev": 0.0002593984474151631,
"stdev_": 0.5912493089316151
},
"throughput": {
"unit": "samples/s",
"value": 136.75856154651456
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_text-generation_openai-community/gpt2 | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model_type": "gpt2",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1350.053888,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.35609701538085936,
0.044365825653076174,
0.04398284912109375,
0.043514881134033206,
0.04412108612060547
],
"count": 5,
"total": 0.532081657409668,
"mean": 0.1064163314819336,
"p50": 0.04412108612060547,
"p90": 0.2314045394897461,
"p95": 0.29375077743530265,
"p99": 0.34362776779174803,
"stdev": 0.12484064959463413,
"stdev_": 117.3134309895173
},
"throughput": {
"unit": "samples/s",
"value": 93.97053873913808
},
"energy": {
"unit": "kWh",
"cpu": 0.00001046066729236246,
"ram": 0.000005713810620613953,
"gpu": 0.000015273067774000063,
"total": 0.00003144754568697648
},
"efficiency": {
"unit": "samples/kWh",
"value": 317989.8393196817
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1350.053888,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.35609701538085936,
0.044365825653076174
],
"count": 2,
"total": 0.40046284103393553,
"mean": 0.20023142051696777,
"p50": 0.20023142051696777,
"p90": 0.32492389640808106,
"p95": 0.3405104558944702,
"p99": 0.35297970348358154,
"stdev": 0.1558655948638916,
"stdev_": 77.84272541315933
},
"throughput": {
"unit": "samples/s",
"value": 19.976884695082294
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1350.053888,
"max_global_vram": 3566.731264,
"max_process_vram": 0,
"max_reserved": 2910.846976,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"values": [
0.04398284912109375,
0.043514881134033206,
0.04412108612060547
],
"count": 3,
"total": 0.13161881637573242,
"mean": 0.04387293879191081,
"p50": 0.04398284912109375,
"p90": 0.04409343872070312,
"p95": 0.0441072624206543,
"p99": 0.044118321380615236,
"stdev": 0.0002593984474151631,
"stdev_": 0.5912493089316151
},
"throughput": {
"unit": "samples/s",
"value": 136.75856154651456
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"model": "openai-community/gpt2",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 5,
"total": 0.8139294586181639,
"mean": 0.16278589172363278,
"stdev": 0.2273662362653359,
"p50": 0.04927385711669922,
"p90": 0.3902293930053711,
"p95": 0.5038737297058105,
"p99": 0.594789199066162,
"values": [
0.61751806640625,
0.049296382904052735,
0.04860006332397461,
0.0492410888671875,
0.04927385711669922
]
},
"throughput": {
"unit": "samples/s",
"value": 61.4303849929289
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.6668144493103026,
"mean": 0.3334072246551513,
"stdev": 0.2841108417510986,
"p50": 0.3334072246551513,
"p90": 0.5606958980560303,
"p95": 0.5891069822311401,
"p99": 0.6118358495712279,
"values": [
0.61751806640625,
0.049296382904052735
]
},
"throughput": {
"unit": "samples/s",
"value": 11.99734050195603
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1107.173376,
"max_global_vram": 3563.585536,
"max_process_vram": 0,
"max_reserved": 2915.04128,
"max_allocated": 2523.776
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.1471150093078613,
"mean": 0.04903833643595377,
"stdev": 0.00031019448743967263,
"p50": 0.0492410888671875,
"p90": 0.04926730346679687,
"p95": 0.049270580291748044,
"p99": 0.04927320175170898,
"values": [
0.04860006332397461,
0.0492410888671875,
0.04927385711669922
]
},
"throughput": {
"unit": "samples/s",
"value": 122.35325331307405
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
{
"name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
},
"print_report": true,
"log_report": true
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1386.917888,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.4221747131347656,
0.09074073791503906,
0.08792063903808593,
0.08825856018066407,
0.08857084655761718
],
"count": 5,
"total": 0.7776654968261717,
"mean": 0.15553309936523435,
"p50": 0.08857084655761718,
"p90": 0.289601123046875,
"p95": 0.35588791809082027,
"p99": 0.4089173541259766,
"stdev": 0.1333244553793267,
"stdev_": 85.72095323982731
},
"throughput": {
"unit": "samples/s",
"value": 64.2949959899999
},
"energy": {
"unit": "kWh",
"cpu": 0.000013409542672222587,
"ram": 0.000007298270686959867,
"gpu": 0.00002345001876000001,
"total": 0.00004415783211918246
},
"efficiency": {
"unit": "samples/kWh",
"value": 226460.39264359474
}
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1386.917888,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.4221747131347656,
0.09074073791503906
],
"count": 2,
"total": 0.5129154510498046,
"mean": 0.2564577255249023,
"p50": 0.2564577255249023,
"p90": 0.38903131561279297,
"p95": 0.40560301437377927,
"p99": 0.4188603733825684,
"stdev": 0.16571698760986328,
"stdev_": 64.61766252924676
},
"throughput": {
"unit": "samples/s",
"value": 15.597112513623989
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1386.917888,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.08792063903808593,
0.08825856018066407,
0.08857084655761718
],
"count": 3,
"total": 0.2647500457763672,
"mean": 0.08825001525878906,
"p50": 0.08825856018066407,
"p90": 0.08850838928222655,
"p95": 0.08853961791992188,
"p99": 0.08856460083007812,
"stdev": 0.00026551486635231693,
"stdev_": 0.3008666520608602
},
"throughput": {
"unit": "samples/s",
"value": 67.98865679972155
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |
null | null | cuda_training_transformers_token-classification_microsoft/deberta-v3-base | {
"name": "pytorch",
"version": "2.4.1+cu124",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model_type": "deberta-v2",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"evaluation_strategy": "no",
"eval_strategy": "no",
"save_strategy": "no",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": true
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.261056,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.225-213.878.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.12",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.5.0.dev0",
"optimum_benchmark_commit": null,
"transformers_version": "4.45.2",
"transformers_commit": null,
"accelerate_version": "1.0.1",
"accelerate_commit": null,
"diffusers_version": "0.30.3",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.9",
"timm_commit": null,
"peft_version": "0.13.2",
"peft_commit": null
} | true | true | null | null | null |
null | null | null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 1386.917888,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.4221747131347656,
0.09074073791503906,
0.08792063903808593,
0.08825856018066407,
0.08857084655761718
],
"count": 5,
"total": 0.7776654968261717,
"mean": 0.15553309936523435,
"p50": 0.08857084655761718,
"p90": 0.289601123046875,
"p95": 0.35588791809082027,
"p99": 0.4089173541259766,
"stdev": 0.1333244553793267,
"stdev_": 85.72095323982731
},
"throughput": {
"unit": "samples/s",
"value": 64.2949959899999
},
"energy": {
"unit": "kWh",
"cpu": 0.000013409542672222587,
"ram": 0.000007298270686959867,
"gpu": 0.00002345001876000001,
"total": 0.00004415783211918246
},
"efficiency": {
"unit": "samples/kWh",
"value": 226460.39264359474
}
} | {
"memory": {
"unit": "MB",
"max_ram": 1386.917888,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.4221747131347656,
0.09074073791503906
],
"count": 2,
"total": 0.5129154510498046,
"mean": 0.2564577255249023,
"p50": 0.2564577255249023,
"p90": 0.38903131561279297,
"p95": 0.40560301437377927,
"p99": 0.4188603733825684,
"stdev": 0.16571698760986328,
"stdev_": 64.61766252924676
},
"throughput": {
"unit": "samples/s",
"value": 15.597112513623989
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 1386.917888,
"max_global_vram": 4604.821504,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"values": [
0.08792063903808593,
0.08825856018066407,
0.08857084655761718
],
"count": 3,
"total": 0.2647500457763672,
"mean": 0.08825001525878906,
"p50": 0.08825856018066407,
"p90": 0.08850838928222655,
"p95": 0.08853961791992188,
"p99": 0.08856460083007812,
"stdev": 0.00026551486635231693,
"stdev_": 0.3008666520608602
},
"throughput": {
"unit": "samples/s",
"value": 67.98865679972155
},
"energy": null,
"efficiency": null
} |
{
"name": "cuda_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.2.2",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"model": "microsoft/deberta-v3-base",
"library": "transformers",
"device": "cuda",
"device_ids": "0",
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": true,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7R32",
"cpu_count": 16,
"cpu_ram_mb": 66697.29792,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"gpu": [
"NVIDIA A10G"
],
"gpu_count": 1,
"gpu_vram_mb": 24146608128,
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": null,
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 5,
"total": 1.0522736740112304,
"mean": 0.21045473480224608,
"stdev": 0.2521626361507063,
"p50": 0.08427519989013672,
"p90": 0.46303764648437507,
"p95": 0.5889081359863281,
"p99": 0.6896045275878907,
"values": [
0.7147786254882813,
0.08542617797851562,
0.08360550689697266,
0.08418816375732421,
0.08427519989013672
]
},
"throughput": {
"unit": "samples/s",
"value": 47.51615595342393
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 2,
"total": 0.8002048034667969,
"mean": 0.40010240173339845,
"stdev": 0.3146762237548828,
"p50": 0.40010240173339845,
"p90": 0.6518433807373047,
"p95": 0.683311003112793,
"p99": 0.7084851010131836,
"values": [
0.7147786254882813,
0.08542617797851562
]
},
"throughput": {
"unit": "samples/s",
"value": 9.997440611879489
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 1140.764672,
"max_global_vram": 4597.481472,
"max_process_vram": 0,
"max_reserved": 3948.937216,
"max_allocated": 3702.95552
},
"latency": {
"unit": "s",
"count": 3,
"total": 0.2520688705444336,
"mean": 0.08402295684814454,
"stdev": 0.0002973125946472148,
"p50": 0.08418816375732421,
"p90": 0.08425779266357422,
"p95": 0.08426649627685547,
"p99": 0.08427345916748047,
"values": [
0.08360550689697266,
0.08418816375732421,
0.08427519989013672
]
},
"throughput": {
"unit": "samples/s",
"value": 71.40905563278207
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null | null | null |