{ "config": { "name": "cpu_inference_transformers_text-generation_openai-community/gpt2", "backend": { "name": "pytorch", "version": "2.5.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-generation", "library": "transformers", "model_type": "gpt2", "model": "openai-community/gpt2", "processor": "openai-community/gpt2", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "inference", "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", "iterations": 1, "duration": 1, "warmup_runs": 1, "input_shapes": { "batch_size": 2, "sequence_length": 16 }, "new_tokens": null, "memory": true, "latency": true, "energy": true, "forward_kwargs": {}, "generate_kwargs": { "max_new_tokens": 2, "min_new_tokens": 2 }, "call_kwargs": { "num_inference_steps": 2 } }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": null, "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7763 64-Core Processor", "cpu_count": 4, "cpu_ram_mb": 16757.342208, "system": "Linux", "machine": "x86_64", "platform": "Linux-6.5.0-1025-azure-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.15", "optimum_benchmark_version": "0.5.0.dev0", "optimum_benchmark_commit": "04bb8676c2e912cf262caaa2d4796a0191a83273", "transformers_version": "4.46.3", "transformers_commit": null, "accelerate_version": "1.1.1", "accelerate_commit": null, "diffusers_version": "0.31.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.11", "timm_commit": null, "peft_version": null, "peft_commit": null }, "print_report": true, "log_report": true }, "report": { "load": { "memory": { "unit": "MB", "max_ram": 1130.233856, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 1.7327888100000166 ], "count": 1, "total": 1.7327888100000166, "mean": 1.7327888100000166, "p50": 1.7327888100000166, "p90": 1.7327888100000166, "p95": 1.7327888100000166, "p99": 1.7327888100000166, "stdev": 0, "stdev_": 0 }, "throughput": null, "energy": null, "efficiency": null }, "prefill": { "memory": { "unit": "MB", "max_ram": 1012.641792, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.15366975499998148, 0.16072087899999588, 0.11341367500000388, 0.10321070399999144, 0.10557986899999605, 0.10479721300001188 ], "count": 6, "total": 0.7413920949999806, "mean": 0.12356534916666344, "p50": 0.10949677199999996, "p90": 0.15719531699998868, "p95": 0.15895809799999228, "p99": 0.16036832279999516, "stdev": 0.024082833263604584, "stdev_": 19.489956873849763 }, "throughput": { "unit": "tokens/s", "value": 258.9722783596782 }, "energy": { "unit": "kWh", "cpu": 4.218644694444484e-06, "ram": 1.7630452732304486e-07, "gpu": 0.0, "total": 4.3949492217675285e-06 }, "efficiency": { "unit": "tokens/kWh", "value": 7281085.260668945 } }, "decode": { "memory": { "unit": "MB", "max_ram": 1020.64128, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "values": [ 0.05179725900001131, 0.052174370000017234, 0.04633807199999751, 0.044985338000003594, 0.04457143900000915, 0.04629236700000661 ], "count": 6, "total": 0.2861588450000454, "mean": 0.0476931408333409, "p50": 0.04631521950000206, "p90": 0.05198581450001427, "p95": 0.05208009225001575, "p99": 0.05215551445001694, "stdev": 0.0031037825891959037, "stdev_": 6.507817549785145 }, "throughput": { "unit": "tokens/s", "value": 41.93475130918318 }, "energy": { "unit": "kWh", "cpu": 1.6928337666666454e-06, "ram": 7.074798652428004e-08, "gpu": 0.0, "total": 1.7635817531909286e-06 }, "efficiency": { "unit": "tokens/kWh", "value": 1134055.7342359147 } }, "per_token": { "memory": null, "latency": { "unit": "s", "values": [ 0.05067520299999728, 0.05078066000001513, 0.04527658899999665, 0.043886185000019395, 0.043664963999987094, 0.04519753200000309 ], "count": 6, "total": 0.27948113300001864, "mean": 0.04658018883333644, "p50": 0.04523706049999987, "p90": 0.05072793150000621, "p95": 0.05075429575001067, "p99": 0.05077538715001424, "stdev": 0.002993893896316374, "stdev_": 6.427397508044683 }, "throughput": { "unit": "tokens/s", "value": 42.93670871872127 }, "energy": null, "efficiency": null } } }