Dataset Preview
View in Dataset Viewer
Viewer
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed
Error code: DatasetGenerationError Exception: DatasetGenerationError Message: An error occurred while generating the dataset Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2011, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 583, in write_table self._build_writer(inferred_schema=pa_table.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 404, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1016, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 1869, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2027, in _prepare_split_single num_examples, num_bytes = writer.finalize() File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 602, in finalize self._build_writer(self.schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 404, in _build_writer self.pa_writer = self._WRITER_CLASS(self.stream, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/pyarrow/parquet/core.py", line 1016, in __init__ self.writer = _parquet.ParquetWriter( File "pyarrow/_parquet.pyx", line 1869, in pyarrow._parquet.ParquetWriter.__cinit__ File "pyarrow/error.pxi", line 154, in pyarrow.lib.pyarrow_internal_check_status File "pyarrow/error.pxi", line 91, in pyarrow.lib.check_status pyarrow.lib.ArrowNotImplementedError: Cannot write struct type 'model_kwargs' with no child field to Parquet. Consider adding a dummy child field. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1324, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 938, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1027, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1122, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1882, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 2038, in _prepare_split_single raise DatasetGenerationError("An error occurred while generating the dataset") from e datasets.exceptions.DatasetGenerationError: An error occurred while generating the dataset
Need help to make the dataset viewer work? Open a discussion for direct support.
config
dict | report
dict | name
string | backend
dict | scenario
dict | launcher
dict | environment
dict | overall
dict | warmup
dict | train
dict |
---|---|---|---|---|---|---|---|---|---|
{
"name": "cpu_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2467.688448,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 2.948843576999991,
"mean": 0.5897687153999982,
"stdev": 0.04447695658495729,
"p50": 0.5693590459999882,
"p90": 0.6359335120000082,
"p95": 0.6571748630000229,
"p99": 0.6741679438000346,
"values": [
0.6784162140000376,
0.5693590459999882,
0.5722094589999642,
0.5609958910000046,
0.5678629669999964
]
},
"throughput": {
"unit": "samples/s",
"value": 16.95579934791507
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2467.688448,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.2477752600000258,
"mean": 0.6238876300000129,
"stdev": 0.054528584000024694,
"p50": 0.6238876300000129,
"p90": 0.6675104972000326,
"p95": 0.6729633556000352,
"p99": 0.6773256423200371,
"values": [
0.6784162140000376,
0.5693590459999882
]
},
"throughput": {
"unit": "samples/s",
"value": 6.411410977967166
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2467.688448,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7010683169999652,
"mean": 0.5670227723333218,
"stdev": 0.004616309646666637,
"p50": 0.5678629669999964,
"p90": 0.5713401605999706,
"p95": 0.5717748097999674,
"p99": 0.5721225291599649,
"values": [
0.5722094589999642,
0.5609958910000046,
0.5678629669999964
]
},
"throughput": {
"unit": "samples/s",
"value": 10.5815855954246
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_fill-mask_google-bert/bert-base-uncased | {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"library": "transformers",
"model": "google-bert/bert-base-uncased",
"processor": "google-bert/bert-base-uncased",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2467.688448,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 2.948843576999991,
"mean": 0.5897687153999982,
"stdev": 0.04447695658495729,
"p50": 0.5693590459999882,
"p90": 0.6359335120000082,
"p95": 0.6571748630000229,
"p99": 0.6741679438000346,
"values": [
0.6784162140000376,
0.5693590459999882,
0.5722094589999642,
0.5609958910000046,
0.5678629669999964
]
},
"throughput": {
"unit": "samples/s",
"value": 16.95579934791507
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2467.688448,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.2477752600000258,
"mean": 0.6238876300000129,
"stdev": 0.054528584000024694,
"p50": 0.6238876300000129,
"p90": 0.6675104972000326,
"p95": 0.6729633556000352,
"p99": 0.6773256423200371,
"values": [
0.6784162140000376,
0.5693590459999882
]
},
"throughput": {
"unit": "samples/s",
"value": 6.411410977967166
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2467.688448,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7010683169999652,
"mean": 0.5670227723333218,
"stdev": 0.004616309646666637,
"p50": 0.5678629669999964,
"p90": 0.5713401605999706,
"p95": 0.5717748097999674,
"p99": 0.5721225291599649,
"values": [
0.5722094589999642,
0.5609958910000046,
0.5678629669999964
]
},
"throughput": {
"unit": "samples/s",
"value": 10.5815855954246
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_fill-mask_google-bert/bert-base-uncased",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "fill-mask",
"model": "google-bert/bert-base-uncased",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 2.738775005999969,
"mean": 0.5477550011999938,
"stdev": 0.03693447784258994,
"p50": 0.5307143729999666,
"p90": 0.5856752317999963,
"p95": 0.6036043034000045,
"p99": 0.6179475606800111,
"values": [
0.6215333750000127,
0.5307143729999666,
0.5270229210000252,
0.5276163199999928,
0.5318880169999716
]
},
"throughput": {
"unit": "samples/s",
"value": 18.25633719106628
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.1522477479999793,
"mean": 0.5761238739999897,
"stdev": 0.045409501000023056,
"p50": 0.5761238739999897,
"p90": 0.612451474800008,
"p95": 0.6169924249000104,
"p99": 0.6206251849800123,
"values": [
0.6215333750000127,
0.5307143729999666
]
},
"throughput": {
"unit": "samples/s",
"value": 6.942951300088064
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2488.782848,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.5865272579999896,
"mean": 0.5288424193333299,
"stdev": 0.0021671455040491463,
"p50": 0.5276163199999928,
"p90": 0.5310336775999758,
"p95": 0.5314608472999737,
"p99": 0.531802583059972,
"values": [
0.5270229210000252,
0.5276163199999928,
0.5318880169999716
]
},
"throughput": {
"unit": "samples/s",
"value": 11.345534663357242
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2437.98016,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 7.749667021000164,
"mean": 1.5499334042000328,
"stdev": 0.059876459519418744,
"p50": 1.5198368840000285,
"p90": 1.611866286400027,
"p95": 1.6407070542000268,
"p99": 1.6637796684400268,
"values": [
1.6695478220000268,
1.5176263130000507,
1.5173120190000304,
1.5253439830000275,
1.5198368840000285
]
},
"throughput": {
"unit": "samples/s",
"value": 6.451890108892324
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2437.98016,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 3.1871741350000775,
"mean": 1.5935870675000388,
"stdev": 0.07596075449998807,
"p50": 1.5935870675000388,
"p90": 1.6543556711000291,
"p95": 1.661951746550028,
"p99": 1.668028606910027,
"values": [
1.6695478220000268,
1.5176263130000507
]
},
"throughput": {
"unit": "samples/s",
"value": 2.5100605304704526
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2437.98016,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 4.562492886000086,
"mean": 1.5208309620000289,
"stdev": 0.0033535309478096545,
"p50": 1.5198368840000285,
"p90": 1.5242425632000276,
"p95": 1.5247932731000275,
"p99": 1.5252338410200275,
"values": [
1.5173120190000304,
1.5253439830000275,
1.5198368840000285
]
},
"throughput": {
"unit": "samples/s",
"value": 3.9452116309556606
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_image-classification_google/vit-base-patch16-224 | {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"library": "transformers",
"model": "google/vit-base-patch16-224",
"processor": "google/vit-base-patch16-224",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2437.98016,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 7.749667021000164,
"mean": 1.5499334042000328,
"stdev": 0.059876459519418744,
"p50": 1.5198368840000285,
"p90": 1.611866286400027,
"p95": 1.6407070542000268,
"p99": 1.6637796684400268,
"values": [
1.6695478220000268,
1.5176263130000507,
1.5173120190000304,
1.5253439830000275,
1.5198368840000285
]
},
"throughput": {
"unit": "samples/s",
"value": 6.451890108892324
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2437.98016,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 3.1871741350000775,
"mean": 1.5935870675000388,
"stdev": 0.07596075449998807,
"p50": 1.5935870675000388,
"p90": 1.6543556711000291,
"p95": 1.661951746550028,
"p99": 1.668028606910027,
"values": [
1.6695478220000268,
1.5176263130000507
]
},
"throughput": {
"unit": "samples/s",
"value": 2.5100605304704526
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2437.98016,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 4.562492886000086,
"mean": 1.5208309620000289,
"stdev": 0.0033535309478096545,
"p50": 1.5198368840000285,
"p90": 1.5242425632000276,
"p95": 1.5247932731000275,
"p99": 1.5252338410200275,
"values": [
1.5173120190000304,
1.5253439830000275,
1.5198368840000285
]
},
"throughput": {
"unit": "samples/s",
"value": 3.9452116309556606
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_image-classification_google/vit-base-patch16-224",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "image-classification",
"model": "google/vit-base-patch16-224",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 7.2970974209999895,
"mean": 1.459419484199998,
"stdev": 0.05210139006345095,
"p50": 1.4401334369999859,
"p90": 1.521250764199999,
"p95": 1.5379663595999886,
"p99": 1.5513388359199802,
"values": [
1.5546819549999782,
1.4241662819999874,
1.4401334369999859,
1.4711039780000306,
1.4070117690000075
]
},
"throughput": {
"unit": "samples/s",
"value": 6.852039532336137
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.9788482369999656,
"mean": 1.4894241184999828,
"stdev": 0.06525783649999539,
"p50": 1.4894241184999828,
"p90": 1.541630387699979,
"p95": 1.5481561713499787,
"p99": 1.5533767982699782,
"values": [
1.5546819549999782,
1.4241662819999874
]
},
"throughput": {
"unit": "samples/s",
"value": 2.6856017371522416
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2442.985472,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 4.318249184000024,
"mean": 1.4394163946666747,
"stdev": 0.02617044676610726,
"p50": 1.4401334369999859,
"p90": 1.4649098698000216,
"p95": 1.4680069239000262,
"p99": 1.4704845671800297,
"values": [
1.4401334369999859,
1.4711039780000306,
1.4070117690000075
]
},
"throughput": {
"unit": "samples/s",
"value": 4.1683560241714614
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2833.154048,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.8069755459999897,
"mean": 0.761395109199998,
"stdev": 0.049894389234311276,
"p50": 0.7365560349999782,
"p90": 0.8158855796000012,
"p95": 0.8380173077999984,
"p99": 0.8557226903599963,
"values": [
0.8601490359999957,
0.7494903950000094,
0.732995982000034,
0.7277840979999723,
0.7365560349999782
]
},
"throughput": {
"unit": "samples/s",
"value": 13.133785440921804
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2833.154048,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.609639431000005,
"mean": 0.8048197155000025,
"stdev": 0.05532932049999317,
"p50": 0.8048197155000025,
"p90": 0.849083171899997,
"p95": 0.8546161039499964,
"p99": 0.8590424495899959,
"values": [
0.8601490359999957,
0.7494903950000094
]
},
"throughput": {
"unit": "samples/s",
"value": 4.970057173009186
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2833.154048,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.1973361149999846,
"mean": 0.7324453716666616,
"stdev": 0.003602230650345284,
"p50": 0.732995982000034,
"p90": 0.7358440243999894,
"p95": 0.7362000296999838,
"p99": 0.7364848339399793,
"values": [
0.732995982000034,
0.7277840979999723,
0.7365560349999782
]
},
"throughput": {
"unit": "samples/s",
"value": 8.191737202662837
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_multiple-choice_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"library": "transformers",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2833.154048,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.8069755459999897,
"mean": 0.761395109199998,
"stdev": 0.049894389234311276,
"p50": 0.7365560349999782,
"p90": 0.8158855796000012,
"p95": 0.8380173077999984,
"p99": 0.8557226903599963,
"values": [
0.8601490359999957,
0.7494903950000094,
0.732995982000034,
0.7277840979999723,
0.7365560349999782
]
},
"throughput": {
"unit": "samples/s",
"value": 13.133785440921804
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2833.154048,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.609639431000005,
"mean": 0.8048197155000025,
"stdev": 0.05532932049999317,
"p50": 0.8048197155000025,
"p90": 0.849083171899997,
"p95": 0.8546161039499964,
"p99": 0.8590424495899959,
"values": [
0.8601490359999957,
0.7494903950000094
]
},
"throughput": {
"unit": "samples/s",
"value": 4.970057173009186
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2833.154048,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.1973361149999846,
"mean": 0.7324453716666616,
"stdev": 0.003602230650345284,
"p50": 0.732995982000034,
"p90": 0.7358440243999894,
"p95": 0.7362000296999838,
"p99": 0.7364848339399793,
"values": [
0.732995982000034,
0.7277840979999723,
0.7365560349999782
]
},
"throughput": {
"unit": "samples/s",
"value": 8.191737202662837
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_multiple-choice_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "multiple-choice",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.581090587999995,
"mean": 0.716218117599999,
"stdev": 0.043372798377969854,
"p50": 0.697155070000008,
"p90": 0.7645524997999928,
"p95": 0.7826806403999967,
"p99": 0.7971831528799999,
"values": [
0.8008087810000006,
0.710168077999981,
0.6928677590000234,
0.697155070000008,
0.680090899999982
]
},
"throughput": {
"unit": "samples/s",
"value": 13.962227084549829
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.5109768589999817,
"mean": 0.7554884294999908,
"stdev": 0.04532035150000979,
"p50": 0.7554884294999908,
"p90": 0.7917447106999986,
"p95": 0.7962767458499996,
"p99": 0.7999023739700004,
"values": [
0.8008087810000006,
0.710168077999981
]
},
"throughput": {
"unit": "samples/s",
"value": 5.294588035778844
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2845.749248,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 2.0701137290000133,
"mean": 0.6900379096666711,
"stdev": 0.007248103654729414,
"p50": 0.6928677590000234,
"p90": 0.696297607800011,
"p95": 0.6967263389000096,
"p99": 0.6970693237800083,
"values": [
0.6928677590000234,
0.697155070000008,
0.680090899999982
]
},
"throughput": {
"unit": "samples/s",
"value": 8.695174447587021
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2832.50688,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.11779525299994,
"mean": 0.6235590505999881,
"stdev": 0.049603432971682564,
"p50": 0.5995712669999875,
"p90": 0.6744295051999984,
"p95": 0.698530285600009,
"p99": 0.7178109099200174,
"values": [
0.7226310660000195,
0.5995712669999875,
0.5993030899999781,
0.5941626659999883,
0.6021271639999668
]
},
"throughput": {
"unit": "samples/s",
"value": 16.03697354785887
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2832.50688,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.322202333000007,
"mean": 0.6611011665000035,
"stdev": 0.06152989950001597,
"p50": 0.6611011665000035,
"p90": 0.7103250861000163,
"p95": 0.7164780760500179,
"p99": 0.7214004680100191,
"values": [
0.7226310660000195,
0.5995712669999875
]
},
"throughput": {
"unit": "samples/s",
"value": 6.050511181483415
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2832.50688,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7955929199999332,
"mean": 0.5985309733333111,
"stdev": 0.0032970117994050186,
"p50": 0.5993030899999781,
"p90": 0.6015623491999691,
"p95": 0.601844756599968,
"p99": 0.602070682519967,
"values": [
0.5993030899999781,
0.5941626659999883,
0.6021271639999668
]
},
"throughput": {
"unit": "samples/s",
"value": 10.024543870445129
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_text-classification_FacebookAI/roberta-base | {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"library": "transformers",
"model": "FacebookAI/roberta-base",
"processor": "FacebookAI/roberta-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2832.50688,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.11779525299994,
"mean": 0.6235590505999881,
"stdev": 0.049603432971682564,
"p50": 0.5995712669999875,
"p90": 0.6744295051999984,
"p95": 0.698530285600009,
"p99": 0.7178109099200174,
"values": [
0.7226310660000195,
0.5995712669999875,
0.5993030899999781,
0.5941626659999883,
0.6021271639999668
]
},
"throughput": {
"unit": "samples/s",
"value": 16.03697354785887
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2832.50688,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.322202333000007,
"mean": 0.6611011665000035,
"stdev": 0.06152989950001597,
"p50": 0.6611011665000035,
"p90": 0.7103250861000163,
"p95": 0.7164780760500179,
"p99": 0.7214004680100191,
"values": [
0.7226310660000195,
0.5995712669999875
]
},
"throughput": {
"unit": "samples/s",
"value": 6.050511181483415
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2832.50688,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7955929199999332,
"mean": 0.5985309733333111,
"stdev": 0.0032970117994050186,
"p50": 0.5993030899999781,
"p90": 0.6015623491999691,
"p95": 0.601844756599968,
"p99": 0.602070682519967,
"values": [
0.5993030899999781,
0.5941626659999883,
0.6021271639999668
]
},
"throughput": {
"unit": "samples/s",
"value": 10.024543870445129
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_text-classification_FacebookAI/roberta-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-classification",
"model": "FacebookAI/roberta-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 2.882509665999976,
"mean": 0.5765019331999952,
"stdev": 0.04978939696949581,
"p50": 0.5569985249999831,
"p90": 0.6300386333999881,
"p95": 0.6525100941999881,
"p99": 0.670487262839988,
"values": [
0.674981554999988,
0.5569985249999831,
0.5424031590000027,
0.5626242509999884,
0.5455021760000136
]
},
"throughput": {
"unit": "samples/s",
"value": 17.34599560576128
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.2319800799999712,
"mean": 0.6159900399999856,
"stdev": 0.05899151500000244,
"p50": 0.6159900399999856,
"p90": 0.6631832519999875,
"p95": 0.6690824034999878,
"p99": 0.673801724699988,
"values": [
0.674981554999988,
0.5569985249999831
]
},
"throughput": {
"unit": "samples/s",
"value": 6.493611487614465
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2826.752,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.6505295860000047,
"mean": 0.5501765286666682,
"stdev": 0.00889233078021606,
"p50": 0.5455021760000136,
"p90": 0.5591998359999935,
"p95": 0.5609120434999909,
"p99": 0.5622818094999888,
"values": [
0.5424031590000027,
0.5626242509999884,
0.5455021760000136
]
},
"throughput": {
"unit": "samples/s",
"value": 10.905590637500968
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2826.211328,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.259441523999982,
"mean": 0.6518883047999964,
"stdev": 0.05944377502380996,
"p50": 0.6247093909999819,
"p90": 0.7131526025999904,
"p95": 0.7418334937999873,
"p99": 0.7647782067599849,
"values": [
0.7705143849999843,
0.6217663820000041,
0.6247093909999819,
0.6271099289999995,
0.6153414370000121
]
},
"throughput": {
"unit": "samples/s",
"value": 15.34005124247177
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2826.211328,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.3922807669999884,
"mean": 0.6961403834999942,
"stdev": 0.07437400149999007,
"p50": 0.6961403834999942,
"p90": 0.7556395846999863,
"p95": 0.7630769848499852,
"p99": 0.7690269049699845,
"values": [
0.7705143849999843,
0.6217663820000041
]
},
"throughput": {
"unit": "samples/s",
"value": 5.74596747266571
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2826.211328,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.8671607569999935,
"mean": 0.6223869189999979,
"stdev": 0.0050773850311273735,
"p50": 0.6247093909999819,
"p90": 0.626629821399996,
"p95": 0.6268698751999977,
"p99": 0.6270619182399991,
"values": [
0.6247093909999819,
0.6271099289999995,
0.6153414370000121
]
},
"throughput": {
"unit": "samples/s",
"value": 9.640305438360315
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_text-generation_openai-community/gpt2 | {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"library": "transformers",
"model": "openai-community/gpt2",
"processor": "openai-community/gpt2",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 2826.211328,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.259441523999982,
"mean": 0.6518883047999964,
"stdev": 0.05944377502380996,
"p50": 0.6247093909999819,
"p90": 0.7131526025999904,
"p95": 0.7418334937999873,
"p99": 0.7647782067599849,
"values": [
0.7705143849999843,
0.6217663820000041,
0.6247093909999819,
0.6271099289999995,
0.6153414370000121
]
},
"throughput": {
"unit": "samples/s",
"value": 15.34005124247177
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2826.211328,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.3922807669999884,
"mean": 0.6961403834999942,
"stdev": 0.07437400149999007,
"p50": 0.6961403834999942,
"p90": 0.7556395846999863,
"p95": 0.7630769848499852,
"p99": 0.7690269049699845,
"values": [
0.7705143849999843,
0.6217663820000041
]
},
"throughput": {
"unit": "samples/s",
"value": 5.74596747266571
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 2826.211328,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.8671607569999935,
"mean": 0.6223869189999979,
"stdev": 0.0050773850311273735,
"p50": 0.6247093909999819,
"p90": 0.626629821399996,
"p95": 0.6268698751999977,
"p99": 0.6270619182399991,
"values": [
0.6247093909999819,
0.6271099289999995,
0.6153414370000121
]
},
"throughput": {
"unit": "samples/s",
"value": 9.640305438360315
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_text-generation_openai-community/gpt2",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "text-generation",
"model": "openai-community/gpt2",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 3.1791685380000274,
"mean": 0.6358337076000055,
"stdev": 0.07846941233493662,
"p50": 0.596941285000014,
"p90": 0.7161873142000047,
"p95": 0.7544328206000045,
"p99": 0.7850292257200044,
"values": [
0.7926783270000044,
0.6014507950000052,
0.596941285000014,
0.593667699000008,
0.5944304319999958
]
},
"throughput": {
"unit": "samples/s",
"value": 15.727382616668173
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 1.3941291220000096,
"mean": 0.6970645610000048,
"stdev": 0.0956137659999996,
"p50": 0.6970645610000048,
"p90": 0.7735555738000045,
"p95": 0.7831169504000044,
"p99": 0.7907660516800044,
"values": [
0.7926783270000044,
0.6014507950000052
]
},
"throughput": {
"unit": "samples/s",
"value": 5.738349392288174
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 2827.354112,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 1.7850394160000178,
"mean": 0.5950131386666726,
"stdev": 0.0013985114990352936,
"p50": 0.5944304319999958,
"p90": 0.5964391144000103,
"p95": 0.5966901997000121,
"p99": 0.5968910679400136,
"values": [
0.596941285000014,
0.593667699000008,
0.5944304319999958
]
},
"throughput": {
"unit": "samples/s",
"value": 10.083810944822195
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
{
"name": "cpu_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 4383.37536,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 5.949403631999928,
"mean": 1.1898807263999855,
"stdev": 0.08196632569906208,
"p50": 1.1518058079999491,
"p90": 1.2827086855999936,
"p95": 1.315300377799997,
"p99": 1.3413737315599996,
"values": [
1.3478920700000003,
1.1488080729999979,
1.1518058079999491,
1.1849336089999838,
1.115964071999997
]
},
"throughput": {
"unit": "samples/s",
"value": 8.40420369716825
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 4383.37536,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.496700142999998,
"mean": 1.248350071499999,
"stdev": 0.09954199850000123,
"p50": 1.248350071499999,
"p90": 1.3279836703,
"p95": 1.3379378701500002,
"p99": 1.3459012300300004,
"values": [
1.3478920700000003,
1.1488080729999979
]
},
"throughput": {
"unit": "samples/s",
"value": 3.2042293995254543
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 4383.37536,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 3.45270348899993,
"mean": 1.1509011629999766,
"stdev": 0.028163960959946867,
"p50": 1.1518058079999491,
"p90": 1.1783080487999769,
"p95": 1.1816208288999803,
"p99": 1.1842710529799831,
"values": [
1.1518058079999491,
1.1849336089999838,
1.115964071999997
]
},
"throughput": {
"unit": "samples/s",
"value": 5.2133060534583215
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
null | null | cpu_training_transformers_token-classification_microsoft/deberta-v3-base | {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"library": "transformers",
"model": "microsoft/deberta-v3-base",
"processor": "microsoft/deberta-v3-base",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"model_kwargs": {},
"processor_kwargs": {},
"hub_kwargs": {},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
} | {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
} | {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"numactl": false,
"numactl_kwargs": {},
"start_method": "spawn"
} | {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.354496,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1021-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.1",
"optimum_benchmark_commit": "347e13ca9f7f904f55669603cfb9f0b6c7e8672c",
"transformers_version": "4.41.1",
"transformers_commit": null,
"accelerate_version": "0.30.1",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "1.0.3",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
} | null | null | null |
null | null | null | null | null | null | null | {
"memory": {
"unit": "MB",
"max_ram": 4383.37536,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 5.949403631999928,
"mean": 1.1898807263999855,
"stdev": 0.08196632569906208,
"p50": 1.1518058079999491,
"p90": 1.2827086855999936,
"p95": 1.315300377799997,
"p99": 1.3413737315599996,
"values": [
1.3478920700000003,
1.1488080729999979,
1.1518058079999491,
1.1849336089999838,
1.115964071999997
]
},
"throughput": {
"unit": "samples/s",
"value": 8.40420369716825
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 4383.37536,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.496700142999998,
"mean": 1.248350071499999,
"stdev": 0.09954199850000123,
"p50": 1.248350071499999,
"p90": 1.3279836703,
"p95": 1.3379378701500002,
"p99": 1.3459012300300004,
"values": [
1.3478920700000003,
1.1488080729999979
]
},
"throughput": {
"unit": "samples/s",
"value": 3.2042293995254543
},
"energy": null,
"efficiency": null
} | {
"memory": {
"unit": "MB",
"max_ram": 4383.37536,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 3.45270348899993,
"mean": 1.1509011629999766,
"stdev": 0.028163960959946867,
"p50": 1.1518058079999491,
"p90": 1.1783080487999769,
"p95": 1.1816208288999803,
"p99": 1.1842710529799831,
"values": [
1.1518058079999491,
1.1849336089999838,
1.115964071999997
]
},
"throughput": {
"unit": "samples/s",
"value": 5.2133060534583215
},
"energy": null,
"efficiency": null
} |
{
"name": "cpu_training_transformers_token-classification_microsoft/deberta-v3-base",
"backend": {
"name": "pytorch",
"version": "2.3.0+cpu",
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
"task": "token-classification",
"model": "microsoft/deberta-v3-base",
"library": "transformers",
"device": "cpu",
"device_ids": null,
"seed": 42,
"inter_op_num_threads": null,
"intra_op_num_threads": null,
"hub_kwargs": {
"revision": "main",
"force_download": false,
"local_files_only": false,
"trust_remote_code": false
},
"no_weights": true,
"device_map": null,
"torch_dtype": null,
"eval_mode": true,
"to_bettertransformer": false,
"low_cpu_mem_usage": null,
"attn_implementation": null,
"cache_implementation": null,
"autocast_enabled": false,
"autocast_dtype": null,
"torch_compile": false,
"torch_compile_target": "forward",
"torch_compile_config": {},
"quantization_scheme": null,
"quantization_config": {},
"deepspeed_inference": false,
"deepspeed_inference_config": {},
"peft_type": null,
"peft_config": {}
},
"scenario": {
"name": "training",
"_target_": "optimum_benchmark.scenarios.training.scenario.TrainingScenario",
"max_steps": 5,
"warmup_steps": 2,
"dataset_shapes": {
"dataset_size": 500,
"sequence_length": 16,
"num_choices": 1
},
"training_arguments": {
"per_device_train_batch_size": 2,
"gradient_accumulation_steps": 1,
"output_dir": "./trainer_output",
"do_train": true,
"use_cpu": false,
"max_steps": 5,
"do_eval": false,
"do_predict": false,
"report_to": "none",
"skip_memory_metrics": true,
"ddp_find_unused_parameters": false
},
"latency": true,
"memory": true,
"energy": false
},
"launcher": {
"name": "process",
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
"device_isolation": false,
"device_isolation_action": "error",
"start_method": "spawn"
},
"environment": {
"cpu": " AMD EPYC 7763 64-Core Processor",
"cpu_count": 4,
"cpu_ram_mb": 16757.346304,
"system": "Linux",
"machine": "x86_64",
"platform": "Linux-6.5.0-1018-azure-x86_64-with-glibc2.35",
"processor": "x86_64",
"python_version": "3.10.14",
"optimum_benchmark_version": "0.2.0",
"optimum_benchmark_commit": "2e77e02d1fd3ab0d2e788c3d89c12299219a25e8",
"transformers_version": "4.40.2",
"transformers_commit": null,
"accelerate_version": "0.30.0",
"accelerate_commit": null,
"diffusers_version": "0.27.2",
"diffusers_commit": null,
"optimum_version": null,
"optimum_commit": null,
"timm_version": "0.9.16",
"timm_commit": null,
"peft_version": null,
"peft_commit": null
}
} | {
"overall": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 5,
"total": 5.583659903000068,
"mean": 1.1167319806000138,
"stdev": 0.07853258776030796,
"p50": 1.0731834320000075,
"p90": 1.2025418994000006,
"p95": 1.2374836681999908,
"p99": 1.265437083239983,
"values": [
1.2724254369999812,
1.0731834320000075,
1.0977165930000297,
1.0711078369999996,
1.0692266040000504
]
},
"throughput": {
"unit": "samples/s",
"value": 8.954700119385008
},
"energy": null,
"efficiency": null
},
"warmup": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 2,
"total": 2.3456088689999888,
"mean": 1.1728044344999944,
"stdev": 0.09962100249998684,
"p50": 1.1728044344999944,
"p90": 1.252501236499984,
"p95": 1.2624633367499825,
"p99": 1.2704330169499816,
"values": [
1.2724254369999812,
1.0731834320000075
]
},
"throughput": {
"unit": "samples/s",
"value": 3.4106283045436583
},
"energy": null,
"efficiency": null
},
"train": {
"memory": {
"unit": "MB",
"max_ram": 4374.970368,
"max_global_vram": null,
"max_process_vram": null,
"max_reserved": null,
"max_allocated": null
},
"latency": {
"unit": "s",
"count": 3,
"total": 3.2380510340000797,
"mean": 1.0793503446666932,
"stdev": 0.01300958794585394,
"p50": 1.0711078369999996,
"p90": 1.0923948418000236,
"p95": 1.0950557174000266,
"p99": 1.097184417880029,
"values": [
1.0977165930000297,
1.0711078369999996,
1.0692266040000504
]
},
"throughput": {
"unit": "samples/s",
"value": 5.558899415419021
},
"energy": null,
"efficiency": null
}
} | null | null | null | null | null | null | null | null |
No dataset card yet
New: Create and edit this dataset card directly on the website!
Contribute a Dataset Card- Downloads last month
- 0